[ 547.972429] env[65522]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=65522) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 547.972663] env[65522]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=65522) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 547.973010] env[65522]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=65522) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 547.973365] env[65522]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 548.093557] env[65522]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=65522) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:349}} [ 548.102923] env[65522]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.009s {{(pid=65522) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:372}} [ 548.149258] env[65522]: INFO oslo_service.periodic_task [-] Skipping periodic task _heal_instance_info_cache because its interval is negative [ 548.708303] env[65522]: INFO nova.virt.driver [None req-b2f04035-79de-44db-ac29-709df893271d None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 548.780610] env[65522]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 548.780772] env[65522]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 548.780863] env[65522]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=65522) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 551.976964] env[65522]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-08ae7d6d-ee47-40f3-82b9-dbf50b430bb0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.993432] env[65522]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=65522) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 551.993630] env[65522]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-f6817484-ca39-4411-ab1b-1b8ac454822d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.029682] env[65522]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 55487. [ 552.029918] env[65522]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.249s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 552.030399] env[65522]: INFO nova.virt.vmwareapi.driver [None req-b2f04035-79de-44db-ac29-709df893271d None None] VMware vCenter version: 7.0.3 [ 552.033830] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650a213c-c438-4a42-b9b5-f190e6322bdd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.056180] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ef294a-a4e0-4fd7-8edb-d8cdba7e731e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.062883] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2959208-5d01-4c35-a7e2-fbce39d280f0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.070033] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d157a86-6e8e-473b-a266-243328845b96 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.083561] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cfff7ff-8a67-4426-871a-bb9add8ed0d0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.090496] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92b21e2-1126-4109-b92e-4a59636f5436 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.121676] env[65522]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-22cc376e-80b5-4be9-8982-96d678727a78 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.127757] env[65522]: DEBUG nova.virt.vmwareapi.driver [None req-b2f04035-79de-44db-ac29-709df893271d None None] Extension org.openstack.compute already exists. {{(pid=65522) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:228}} [ 552.130385] env[65522]: INFO nova.compute.provider_config [None req-b2f04035-79de-44db-ac29-709df893271d None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 552.633851] env[65522]: DEBUG nova.context [None req-b2f04035-79de-44db-ac29-709df893271d None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),d07f8fc1-ba2d-41aa-8f4b-940ddd0114c1(cell1) {{(pid=65522) load_cells /opt/stack/nova/nova/context.py:472}} [ 552.634105] env[65522]: INFO nova.utils [None req-b2f04035-79de-44db-ac29-709df893271d None None] The cell worker thread pool MainProcess.cell_worker is initialized [ 552.636280] env[65522]: DEBUG oslo_concurrency.lockutils [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 552.636504] env[65522]: DEBUG oslo_concurrency.lockutils [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 552.637279] env[65522]: DEBUG oslo_concurrency.lockutils [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 552.637780] env[65522]: DEBUG oslo_concurrency.lockutils [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Acquiring lock "d07f8fc1-ba2d-41aa-8f4b-940ddd0114c1" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 552.637969] env[65522]: DEBUG oslo_concurrency.lockutils [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Lock "d07f8fc1-ba2d-41aa-8f4b-940ddd0114c1" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 552.638988] env[65522]: DEBUG oslo_concurrency.lockutils [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Lock "d07f8fc1-ba2d-41aa-8f4b-940ddd0114c1" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 552.660653] env[65522]: INFO dbcounter [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Registered counter for database nova_cell0 [ 552.669224] env[65522]: INFO dbcounter [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Registered counter for database nova_cell1 [ 552.672432] env[65522]: DEBUG oslo_db.sqlalchemy.engines [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=65522) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 552.673089] env[65522]: DEBUG oslo_db.sqlalchemy.engines [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=65522) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 552.678228] env[65522]: ERROR nova.db.main.api [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 552.678228] env[65522]: func(*args, **kwargs) [ 552.678228] env[65522]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 552.678228] env[65522]: self.work.run() [ 552.678228] env[65522]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 552.678228] env[65522]: result = self.fn(*self.args, **self.kwargs) [ 552.678228] env[65522]: File "/opt/stack/nova/nova/utils.py", line 585, in context_wrapper [ 552.678228] env[65522]: return func(*args, **kwargs) [ 552.678228] env[65522]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 552.678228] env[65522]: result = fn(*args, **kwargs) [ 552.678228] env[65522]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 552.678228] env[65522]: return f(*args, **kwargs) [ 552.678228] env[65522]: File "/opt/stack/nova/nova/objects/service.py", line 568, in _db_service_get_minimum_version [ 552.678228] env[65522]: return db.service_get_minimum_version(context, binaries) [ 552.678228] env[65522]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 552.678228] env[65522]: _check_db_access() [ 552.678228] env[65522]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 552.678228] env[65522]: stacktrace = ''.join(traceback.format_stack()) [ 552.678228] env[65522]: [ 552.679280] env[65522]: ERROR nova.db.main.api [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 552.679280] env[65522]: func(*args, **kwargs) [ 552.679280] env[65522]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 552.679280] env[65522]: self.work.run() [ 552.679280] env[65522]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 552.679280] env[65522]: result = self.fn(*self.args, **self.kwargs) [ 552.679280] env[65522]: File "/opt/stack/nova/nova/utils.py", line 585, in context_wrapper [ 552.679280] env[65522]: return func(*args, **kwargs) [ 552.679280] env[65522]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 552.679280] env[65522]: result = fn(*args, **kwargs) [ 552.679280] env[65522]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 552.679280] env[65522]: return f(*args, **kwargs) [ 552.679280] env[65522]: File "/opt/stack/nova/nova/objects/service.py", line 568, in _db_service_get_minimum_version [ 552.679280] env[65522]: return db.service_get_minimum_version(context, binaries) [ 552.679280] env[65522]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 552.679280] env[65522]: _check_db_access() [ 552.679280] env[65522]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 552.679280] env[65522]: stacktrace = ''.join(traceback.format_stack()) [ 552.679280] env[65522]: [ 552.679767] env[65522]: WARNING nova.objects.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 552.679883] env[65522]: WARNING nova.objects.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Failed to get minimum service version for cell d07f8fc1-ba2d-41aa-8f4b-940ddd0114c1 [ 552.680312] env[65522]: DEBUG oslo_concurrency.lockutils [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Acquiring lock "singleton_lock" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.680464] env[65522]: DEBUG oslo_concurrency.lockutils [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Acquired lock "singleton_lock" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 552.680696] env[65522]: DEBUG oslo_concurrency.lockutils [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Releasing lock "singleton_lock" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 552.681021] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Full set of CONF: {{(pid=65522) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/service.py:275}} [ 552.681152] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ******************************************************************************** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2804}} [ 552.681271] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] Configuration options gathered from: {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2805}} [ 552.681394] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 552.681578] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 552.681707] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ================================================================================ {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 552.681899] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] allow_resize_to_same_host = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.682074] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] arq_binding_timeout = 300 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.682200] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] backdoor_port = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.682316] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] backdoor_socket = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.682477] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] block_device_allocate_retries = 60 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.682615] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] block_device_allocate_retries_interval = 3 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.682766] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cell_worker_thread_pool_size = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.682923] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cert = self.pem {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.683087] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.683246] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute_monitors = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.683401] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] config_dir = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.683561] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] config_drive_format = iso9660 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.683684] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.683835] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] config_source = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.683990] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] console_host = devstack {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.684154] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] control_exchange = nova {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.684300] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cpu_allocation_ratio = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.684451] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] daemon = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.684634] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] debug = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.684790] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] default_access_ip_network_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.684945] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] default_availability_zone = nova {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.685110] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] default_ephemeral_format = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.685263] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] default_green_pool_size = 1000 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.685479] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.685630] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] default_schedule_zone = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.685785] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] default_thread_pool_size = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.685924] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] disk_allocation_ratio = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.686091] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] enable_new_services = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.686244] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] flat_injected = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.686390] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] force_config_drive = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.686534] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] force_raw_images = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.686713] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] graceful_shutdown_timeout = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.686864] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] heal_instance_info_cache_interval = -1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.687113] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] host = cpu-1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.687289] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.687487] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] initial_disk_allocation_ratio = 1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.687593] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] initial_ram_allocation_ratio = 1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.687823] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.687983] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] instance_build_timeout = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.688153] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] instance_delete_interval = 300 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.688310] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] instance_format = [instance: %(uuid)s] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.688466] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] instance_name_template = instance-%08x {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.688617] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] instance_usage_audit = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.688780] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] instance_usage_audit_period = month {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.688933] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.689100] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] instances_path = /opt/stack/data/nova/instances {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.689257] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] internal_service_availability_zone = internal {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.689400] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] key = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.689547] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] live_migration_retry_count = 30 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.689702] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] log_color = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.689859] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] log_config_append = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.690022] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.690172] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] log_dir = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.690323] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] log_file = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.690446] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] log_options = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.690596] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] log_rotate_interval = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.690753] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] log_rotate_interval_type = days {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.690910] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] log_rotation_type = none {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.691038] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.691159] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.691314] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.691469] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.691589] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.691740] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] long_rpc_timeout = 1800 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.691886] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] max_concurrent_builds = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.692044] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] max_concurrent_live_migrations = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.692193] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] max_concurrent_snapshots = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.692341] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] max_local_block_devices = 3 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.692489] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] max_logfile_count = 30 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.692638] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] max_logfile_size_mb = 200 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.692787] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] maximum_instance_delete_attempts = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.692934] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] migrate_max_retries = -1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.693099] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] mkisofs_cmd = genisoimage {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.693290] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] my_block_storage_ip = 10.180.1.21 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.693412] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] my_ip = 10.180.1.21 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.693604] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.693760] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] network_allocate_retries = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.693925] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.694090] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] osapi_compute_unique_server_name_scope = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.694242] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] password_length = 12 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.694390] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] periodic_enable = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.694578] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] periodic_fuzzy_delay = 60 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.694730] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] pointer_model = usbtablet {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.694884] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] preallocate_images = none {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.695042] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] publish_errors = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.695164] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] pybasedir = /opt/stack/nova {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.695311] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ram_allocation_ratio = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.695459] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] rate_limit_burst = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.695613] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] rate_limit_except_level = CRITICAL {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.695762] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] rate_limit_interval = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.695906] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] reboot_timeout = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.696060] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] reclaim_instance_interval = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.696206] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] record = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.696360] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] reimage_timeout_per_gb = 60 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.696514] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] report_interval = 120 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.696691] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] rescue_timeout = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.696847] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] reserved_host_cpus = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.696996] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] reserved_host_disk_mb = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.697159] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] reserved_host_memory_mb = 512 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.697310] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] reserved_huge_pages = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.697458] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] resize_confirm_window = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.697608] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] resize_fs_using_block_device = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.697779] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] resume_guests_state_on_host_boot = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.697945] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.698106] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] rpc_response_timeout = 60 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.698257] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] run_external_periodic_tasks = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.698411] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] running_deleted_instance_action = reap {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.698558] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] running_deleted_instance_poll_interval = 1800 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.698707] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] running_deleted_instance_timeout = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.698855] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] scheduler_instance_sync_interval = 120 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.699024] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] service_down_time = 720 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.699184] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] servicegroup_driver = db {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.699330] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] shell_completion = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.699480] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] shelved_offload_time = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.699629] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] shelved_poll_interval = 3600 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.699789] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] shutdown_timeout = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.699944] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] source_is_ipv6 = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.700103] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ssl_only = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.700354] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.700513] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] sync_power_state_interval = 600 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.700664] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] sync_power_state_pool_size = 1000 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.700823] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] syslog_log_facility = LOG_USER {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.700969] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] tempdir = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.701130] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] thread_pool_statistic_period = -1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.701280] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] timeout_nbd = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.701436] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] transport_url = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.701584] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] update_resources_interval = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.701731] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] use_cow_images = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.701876] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] use_journal = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.702031] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] use_json = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.702181] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] use_rootwrap_daemon = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.702327] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] use_stderr = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.702471] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] use_syslog = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.702613] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vcpu_pin_set = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.702768] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plugging_is_fatal = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.702919] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plugging_timeout = 300 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.703083] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] virt_mkfs = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.703233] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] volume_usage_poll_interval = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.703380] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] watch_log_file = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.703533] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] web = /usr/share/spice-html5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 552.703705] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_concurrency.disable_process_locking = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.704302] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.704507] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_brick.lock_path = /opt/stack/data/os_brick {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.704687] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.704847] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.705024] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.705182] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.705342] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.705498] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.705649] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.705810] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.compute_link_prefix = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.705968] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.706211] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.dhcp_domain = novalocal {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.706387] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.enable_instance_password = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.706540] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.glance_link_prefix = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.706725] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.706891] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.707057] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.instance_list_per_project_cells = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.707212] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.list_records_by_skipping_down_cells = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.707363] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.local_metadata_per_cell = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.707610] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.max_limit = 1000 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.707795] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.metadata_cache_expiration = 15 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.707966] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.neutron_default_project_id = default {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.708142] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.response_validation = warn {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.708297] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.use_neutron_default_nets = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.708451] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.708604] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.708759] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.708918] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.709083] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.vendordata_dynamic_targets = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.709238] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.vendordata_jsonfile_path = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.709401] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.709582] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.backend = dogpile.cache.memcached {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.709743] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.backend_argument = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.709892] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.backend_expiration_time = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.710069] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.config_prefix = cache.oslo {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.710226] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.debug_cache_backend = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.710378] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.enable_retry_client = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.710528] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.enable_socket_keepalive = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.710688] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.enabled = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.710841] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.enforce_fips_mode = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.710991] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.expiration_time = 600 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.711168] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.hashclient_dead_timeout = 60.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.711320] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.hashclient_retry_attempts = 2 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.711480] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.hashclient_retry_timeout = 1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.711634] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.memcache_dead_retry = 300 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.711782] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.memcache_password = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.711941] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.712109] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.712263] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.memcache_pool_maxsize = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.712416] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.712572] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.memcache_sasl_enabled = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.712743] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.712899] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.memcache_socket_timeout = 1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.713065] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.memcache_username = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.713226] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.proxies = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.713382] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.redis_db = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.713534] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.redis_password = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.713694] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.redis_sentinel_service_name = mymaster {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.713861] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.714031] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.redis_server = localhost:6379 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.714191] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.redis_socket_timeout = 1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.714342] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.redis_username = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.714518] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.retry_attempts = 2 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.714692] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.retry_delay = 0.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.714850] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.socket_keepalive_count = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.715006] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.socket_keepalive_idle = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.715165] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.socket_keepalive_interval = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.715313] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.tls_allowed_ciphers = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.715458] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.tls_cafile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.715606] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.tls_certfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.715759] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.tls_enabled = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.715906] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cache.tls_keyfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.716077] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.auth_section = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.716241] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.auth_type = password {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.716388] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.cafile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.716552] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.catalog_info = volumev3::publicURL {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.716728] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.certfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.716885] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.collect_timing = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.717048] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.cross_az_attach = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.717204] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.debug = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.717353] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.endpoint_template = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.717507] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.http_retries = 3 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.717670] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.insecure = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.717836] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.keyfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.717999] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.os_region_name = RegionOne {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.718165] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.split_loggers = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.718316] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cinder.timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.718474] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.718623] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute.cpu_dedicated_set = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.718771] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute.cpu_shared_set = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.718925] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute.image_type_exclude_list = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.719086] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.719242] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute.max_concurrent_disk_ops = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.719393] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute.max_disk_devices_to_attach = -1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.719546] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.719704] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.719858] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute.resource_provider_association_refresh = 300 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.720013] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.720173] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute.shutdown_retry_interval = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.720342] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.720509] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] conductor.workers = 2 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.720673] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] console.allowed_origins = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.720825] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] console.ssl_ciphers = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.720984] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] console.ssl_minimum_version = default {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.721154] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] consoleauth.enforce_session_timeout = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.721311] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] consoleauth.token_ttl = 600 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.721470] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.cafile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.721618] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.certfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.721772] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.collect_timing = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.721918] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.connect_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.722077] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.connect_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.722228] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.endpoint_override = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.722377] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.insecure = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.722524] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.keyfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.722671] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.max_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.722817] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.min_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.722962] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.region_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.723126] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.retriable_status_codes = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.723274] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.service_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.723430] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.service_type = accelerator {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.723579] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.split_loggers = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.723731] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.status_code_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.723876] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.status_code_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.724030] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.724199] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.724348] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] cyborg.version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.724525] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.asyncio_connection = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.724689] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.asyncio_slave_connection = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.724851] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.backend = sqlalchemy {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.725025] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.connection = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.725183] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.connection_debug = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.725340] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.connection_parameters = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.725494] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.connection_recycle_time = 3600 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.725644] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.connection_trace = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.725794] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.db_inc_retry_interval = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.725944] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.db_max_retries = 20 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.726106] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.db_max_retry_interval = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.726259] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.db_retry_interval = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.726409] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.max_overflow = 50 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.726589] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.max_pool_size = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.726786] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.max_retries = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.726959] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.727205] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.mysql_wsrep_sync_wait = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.727362] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.pool_timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.727516] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.retry_interval = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.727669] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.slave_connection = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.727820] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.sqlite_synchronous = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.727968] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] database.use_db_reconnect = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.728137] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.asyncio_connection = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.728287] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.asyncio_slave_connection = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.728444] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.backend = sqlalchemy {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.728600] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.connection = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.728754] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.connection_debug = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.728911] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.connection_parameters = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.729079] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.connection_recycle_time = 3600 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.729239] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.connection_trace = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.729391] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.db_inc_retry_interval = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.729542] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.db_max_retries = 20 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.729692] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.db_max_retry_interval = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.729841] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.db_retry_interval = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.729990] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.max_overflow = 50 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.730154] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.max_pool_size = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.730304] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.max_retries = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.730459] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.730605] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.730753] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.pool_timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.730899] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.retry_interval = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.731056] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.slave_connection = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.731209] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] api_database.sqlite_synchronous = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.731370] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] devices.enabled_mdev_types = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.731534] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.731693] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ephemeral_storage_encryption.default_format = luks {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.731847] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ephemeral_storage_encryption.enabled = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.731999] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.732202] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.api_servers = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.732364] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.cafile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.732518] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.certfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.732702] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.collect_timing = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.732828] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.connect_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.732975] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.connect_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.733143] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.debug = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.733298] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.default_trusted_certificate_ids = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.733447] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.enable_certificate_validation = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.733598] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.enable_rbd_download = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.733747] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.endpoint_override = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.733903] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.insecure = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.734064] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.keyfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.734216] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.max_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.734362] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.min_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.734533] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.num_retries = 3 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.734705] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.rbd_ceph_conf = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.734859] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.rbd_connect_timeout = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.735027] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.rbd_pool = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.735191] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.rbd_user = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.735343] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.region_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.735491] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.retriable_status_codes = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.735639] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.service_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.735807] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.service_type = image {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.735960] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.split_loggers = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.736120] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.status_code_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.736300] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.status_code_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.736459] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.736665] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.736822] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.verify_glance_signatures = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.736999] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] glance.version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.737216] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] guestfs.debug = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.737386] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.auth_section = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.737541] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.auth_type = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.737694] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.cafile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.737844] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.certfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.737998] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.collect_timing = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.738164] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.connect_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.738315] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.connect_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.738465] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.endpoint_override = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.738618] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.insecure = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.738767] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.keyfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.738924] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.max_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.739075] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.min_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.739226] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.region_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.739373] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.retriable_status_codes = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.739518] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.service_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.739674] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.service_type = shared-file-system {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.739829] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.share_apply_policy_timeout = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.739980] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.split_loggers = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.740139] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.status_code_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.740287] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.status_code_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.740434] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.740601] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.740752] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] manila.version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.740928] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] mks.enabled = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.741273] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.741453] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] image_cache.manager_interval = 2400 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.741613] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] image_cache.precache_concurrency = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.741775] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] image_cache.remove_unused_base_images = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.741933] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.742104] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.742271] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] image_cache.subdirectory_name = _base {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.742438] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.api_max_retries = 60 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.742593] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.api_retry_interval = 2 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.742745] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.auth_section = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.742899] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.auth_type = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.743067] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.cafile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.743217] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.certfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.743399] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.collect_timing = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.743547] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.conductor_group = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.743695] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.connect_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.743847] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.connect_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.743992] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.endpoint_override = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.744159] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.insecure = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.744309] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.keyfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.745759] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.max_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.745759] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.min_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.745759] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.peer_list = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.745759] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.region_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.745759] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.retriable_status_codes = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.745759] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.serial_console_state_timeout = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.745759] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.service_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.746009] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.service_type = baremetal {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.746009] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.shard = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.746009] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.split_loggers = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.746009] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.status_code_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.746115] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.status_code_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.746258] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.746382] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.746531] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ironic.version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.746729] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.746899] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] key_manager.fixed_key = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.747085] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.747241] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.barbican_api_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.747389] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.barbican_endpoint = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.747552] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.barbican_endpoint_type = public {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.747721] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.barbican_region_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.747877] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.cafile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.748033] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.certfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.748189] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.collect_timing = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.748336] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.insecure = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.748482] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.keyfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.748637] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.number_of_retries = 60 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.748790] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.retry_delay = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.748943] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.send_service_user_token = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.749105] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.split_loggers = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.749255] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.749405] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.verify_ssl = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.749552] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican.verify_ssl_path = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.749707] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican_service_user.auth_section = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.749861] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican_service_user.auth_type = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.750016] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican_service_user.cafile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.750166] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican_service_user.certfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.750319] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican_service_user.collect_timing = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.750470] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican_service_user.insecure = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.750621] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican_service_user.keyfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.750769] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican_service_user.split_loggers = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.750916] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] barbican_service_user.timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.751081] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vault.approle_role_id = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.751232] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vault.approle_secret_id = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.751392] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vault.kv_mountpoint = secret {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.751544] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vault.kv_path = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.751700] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vault.kv_version = 2 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.751850] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vault.namespace = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.752023] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vault.root_token_id = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.752159] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vault.ssl_ca_crt_file = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.752317] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vault.timeout = 60.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.752467] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vault.use_ssl = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.752625] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.752783] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.cafile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.752934] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.certfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.753101] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.collect_timing = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.753254] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.connect_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.753403] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.connect_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.753549] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.endpoint_override = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.753699] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.insecure = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.753848] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.keyfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.753995] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.max_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.754155] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.min_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.754305] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.region_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.754453] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.retriable_status_codes = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.754625] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.service_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.754796] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.service_type = identity {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.754950] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.split_loggers = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.755116] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.status_code_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.755270] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.status_code_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.755421] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.755702] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.755746] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] keystone.version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.755918] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.ceph_mount_options = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.756271] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.756447] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.connection_uri = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.756646] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.cpu_mode = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.756805] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.cpu_model_extra_flags = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.756970] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.cpu_models = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.757145] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.cpu_power_governor_high = performance {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.757303] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.cpu_power_governor_low = powersave {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.757456] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.cpu_power_management = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.757639] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.757802] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.device_detach_attempts = 8 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.757958] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.device_detach_timeout = 20 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.758126] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.disk_cachemodes = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.758278] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.disk_prefix = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.758431] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.enabled_perf_events = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.758588] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.file_backed_memory = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.758744] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.gid_maps = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.758894] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.hw_disk_discard = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.759049] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.hw_machine_type = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.759214] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.images_rbd_ceph_conf = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.759368] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.759516] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.759702] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.images_rbd_glance_store_name = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.759890] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.images_rbd_pool = rbd {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.760065] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.images_type = default {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.760219] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.images_volume_group = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.760371] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.inject_key = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.760521] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.inject_partition = -2 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.760670] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.inject_password = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.760824] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.iscsi_iface = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.760971] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.iser_use_multipath = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.761136] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.live_migration_bandwidth = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.761287] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.761437] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.live_migration_downtime = 500 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.761587] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.761742] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.761966] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.live_migration_inbound_addr = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.762158] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.live_migration_parallel_connections = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.762316] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.762468] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.live_migration_permit_post_copy = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.762616] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.live_migration_scheme = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.762778] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.live_migration_timeout_action = abort {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.762932] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.live_migration_tunnelled = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.763091] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.live_migration_uri = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.763245] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.live_migration_with_native_tls = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.763394] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.max_queues = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.763545] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.763769] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.763923] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.nfs_mount_options = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.764233] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.764398] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.764582] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.num_iser_scan_tries = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.764746] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.num_memory_encrypted_guests = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.764898] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.765061] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.num_pcie_ports = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.765220] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.num_volume_scan_tries = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.765368] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.pmem_namespaces = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.765513] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.quobyte_client_cfg = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.765792] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.765957] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.rbd_connect_timeout = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.766124] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.766278] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.766423] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.rbd_secret_uuid = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.766583] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.rbd_user = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.766752] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.766916] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.remote_filesystem_transport = ssh {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.767078] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.rescue_image_id = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.767228] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.rescue_kernel_id = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.767375] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.rescue_ramdisk_id = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.767532] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.767706] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.rx_queue_size = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.767867] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.smbfs_mount_options = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.768171] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.768333] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.snapshot_compression = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.768485] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.snapshot_image_format = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.768707] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.768865] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.sparse_logical_volumes = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.769044] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.supported_tpm_secret_security = ['user', 'host'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.769200] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.swtpm_enabled = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.769356] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.swtpm_group = tss {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.769512] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.swtpm_user = tss {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.769666] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.sysinfo_serial = unique {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.769812] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.tb_cache_size = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.769956] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.tx_queue_size = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.770117] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.uid_maps = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.770265] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.use_default_aio_mode_for_volumes = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.770411] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.use_virtio_for_bridges = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.770564] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.virt_type = kvm {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.770721] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.volume_clear = zero {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.770871] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.volume_clear_size = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.771029] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.volume_enforce_multipath = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.771181] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.volume_use_multipath = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.771333] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.vzstorage_cache_path = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.771488] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.771641] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.vzstorage_mount_group = qemu {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.771792] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.vzstorage_mount_opts = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.771947] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.772228] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.772393] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.vzstorage_mount_user = stack {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.772549] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.772711] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.auth_section = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.772868] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.auth_type = password {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.773023] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.cafile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.773172] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.certfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.773322] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.collect_timing = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.773467] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.connect_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.773613] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.connect_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.773769] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.default_floating_pool = public {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.773914] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.endpoint_override = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.774072] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.extension_sync_interval = 600 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.774223] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.http_retries = 3 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.774371] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.insecure = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.774539] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.keyfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.774699] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.max_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.774863] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.775014] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.min_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.775184] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.ovs_bridge = br-int {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.775337] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.physnets = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.775493] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.region_name = RegionOne {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.775638] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.retriable_status_codes = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.775794] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.service_metadata_proxy = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.775940] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.service_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.776107] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.service_type = network {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.776259] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.split_loggers = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.776573] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.status_code_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.776573] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.status_code_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.776712] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.776897] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.777044] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] neutron.version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.777205] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] notifications.bdms_in_notifications = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.777353] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] notifications.default_level = INFO {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.777502] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] notifications.include_share_mapping = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.777694] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] notifications.notification_format = unversioned {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.777866] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] notifications.notify_on_state_change = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.778020] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.778186] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] pci.alias = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.778342] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] pci.device_spec = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.778497] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] pci.report_in_placement = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.778657] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.auth_section = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.778818] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.auth_type = password {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.778972] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.779161] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.cafile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.779319] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.certfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.779468] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.collect_timing = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.779615] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.connect_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.779828] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.connect_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.779905] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.default_domain_id = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.780056] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.default_domain_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.780202] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.domain_id = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.780338] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.domain_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.780498] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.endpoint_override = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.780627] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.insecure = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.780759] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.keyfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.780896] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.max_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.781044] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.min_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.781293] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.password = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.781341] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.project_domain_id = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.781480] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.project_domain_name = Default {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.781624] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.project_id = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.781776] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.project_name = service {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.781923] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.region_name = RegionOne {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.782076] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.retriable_status_codes = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784070] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.service_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784070] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.service_type = placement {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784070] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.split_loggers = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784070] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.status_code_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784070] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.status_code_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784070] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.system_scope = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784070] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784275] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.trust_id = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784275] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.user_domain_id = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784275] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.user_domain_name = Default {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784275] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.user_id = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784275] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.username = nova {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784275] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784275] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] placement.version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784443] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.cores = 20 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784443] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.count_usage_from_placement = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784572] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784708] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.injected_file_content_bytes = 10240 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.784859] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.injected_file_path_length = 255 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.785017] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.injected_files = 5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.785170] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.instances = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.785319] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.key_pairs = 100 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.785465] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.metadata_items = 128 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.785619] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.ram = 51200 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.785769] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.recheck_quota = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.785916] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.server_group_members = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.786074] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.server_groups = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.786262] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.unified_limits_resource_list = ['servers'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.786417] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] quota.unified_limits_resource_strategy = require {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.786602] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.786763] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.786908] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] scheduler.image_metadata_prefilter = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.787065] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.787214] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] scheduler.max_attempts = 3 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.787357] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] scheduler.max_placement_results = 1000 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.787499] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.787663] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] scheduler.query_placement_for_image_type_support = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.787812] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.787967] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] scheduler.workers = 2 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.788133] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.788286] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.788449] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.788611] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.788758] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.788903] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.789056] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.789805] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.789805] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.host_subset_size = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.789805] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.789805] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.789805] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.image_props_weight_multiplier = 0.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.790012] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.image_props_weight_setting = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.790082] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.790231] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.isolated_hosts = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.790380] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.isolated_images = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.790543] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.790691] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.790840] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.790987] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.pci_in_placement = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.791151] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.791298] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.791443] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.791588] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.791735] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.791881] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.792040] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.track_instance_changes = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.792288] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.792377] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] metrics.required = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.792511] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] metrics.weight_multiplier = 1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.792661] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.792809] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] metrics.weight_setting = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.793131] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.793297] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] serial_console.enabled = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.793476] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] serial_console.port_range = 10000:20000 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.793619] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.793778] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.793937] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] serial_console.serialproxy_port = 6083 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.794105] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] service_user.auth_section = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.794267] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] service_user.auth_type = password {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.794415] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] service_user.cafile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.794596] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] service_user.certfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.794749] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] service_user.collect_timing = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.794898] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] service_user.insecure = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.795056] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] service_user.keyfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.795220] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] service_user.send_service_user_token = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.795371] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] service_user.split_loggers = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.795518] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] service_user.timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.795673] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] spice.agent_enabled = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.795822] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] spice.enabled = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.796144] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.796344] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.796503] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] spice.html5proxy_port = 6082 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.796692] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] spice.image_compression = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.796851] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] spice.jpeg_compression = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.797008] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] spice.playback_compression = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.797169] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] spice.require_secure = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.797326] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] spice.server_listen = 127.0.0.1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.797483] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.797777] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] spice.spice_direct_proxy_base_url = http://127.0.0.1:13002/nova {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.797944] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] spice.streaming_mode = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.798112] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] spice.zlib_compression = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.798273] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] upgrade_levels.baseapi = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.798435] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] upgrade_levels.compute = auto {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.798586] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] upgrade_levels.conductor = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.798734] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] upgrade_levels.scheduler = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.798889] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vendordata_dynamic_auth.auth_section = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.799049] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vendordata_dynamic_auth.auth_type = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.799198] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vendordata_dynamic_auth.cafile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.799345] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vendordata_dynamic_auth.certfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.799493] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.799639] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vendordata_dynamic_auth.insecure = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.799783] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vendordata_dynamic_auth.keyfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.799930] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.800084] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vendordata_dynamic_auth.timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.800245] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.api_retry_count = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.800391] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.ca_file = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.800547] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.cache_prefix = devstack-image-cache {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.800701] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.cluster_name = testcl1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.800850] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.connection_pool_size = 10 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.800995] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.console_delay_seconds = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.801162] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.datastore_regex = ^datastore.* {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.801366] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.801527] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.host_password = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.801683] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.host_port = 443 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.801840] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.host_username = administrator@vsphere.local {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.801988] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.insecure = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.802150] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.integration_bridge = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.802299] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.maximum_objects = 100 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.802443] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.pbm_default_policy = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.802590] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.pbm_enabled = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.802735] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.pbm_wsdl_location = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.802888] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.803043] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.serial_port_proxy_uri = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.803190] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.serial_port_service_uri = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.803387] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.task_poll_interval = 0.5 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.803506] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.use_linked_clone = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.803732] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.vnc_keymap = en-us {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.803900] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.vnc_port = 5900 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.804064] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vmware.vnc_port_total = 10000 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.804240] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vnc.auth_schemes = ['none'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.804400] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vnc.enabled = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.804699] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.804880] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.805052] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vnc.novncproxy_port = 6080 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.805242] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vnc.server_listen = 127.0.0.1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.805414] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.805567] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vnc.vencrypt_ca_certs = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.805716] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vnc.vencrypt_client_cert = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.805861] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vnc.vencrypt_client_key = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.806028] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.806183] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.disable_deep_image_inspection = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.806332] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.806500] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.806653] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.806815] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.disable_rootwrap = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.806963] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.enable_numa_live_migration = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.807123] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.807276] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.807421] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.807575] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.libvirt_disable_apic = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.807739] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.807889] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.808051] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.808202] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.808348] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.808493] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.808636] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.808781] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.808926] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.809083] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.809251] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.809402] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] wsgi.secure_proxy_ssl_header = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.809551] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] zvm.ca_file = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.809698] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] zvm.cloud_connector_url = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.809996] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.810175] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] zvm.reachable_timeout = 300 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.810345] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.810495] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.810660] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] profiler.connection_string = messaging:// {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.810817] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] profiler.enabled = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.810972] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] profiler.es_doc_type = notification {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.811160] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] profiler.es_scroll_size = 10000 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.811336] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] profiler.es_scroll_time = 2m {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.811487] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] profiler.filter_error_trace = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.811641] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] profiler.hmac_keys = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.811799] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] profiler.sentinel_service_name = mymaster {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.811956] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] profiler.socket_timeout = 0.1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.812119] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] profiler.trace_requests = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.812269] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] profiler.trace_sqlalchemy = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.812432] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] profiler_jaeger.process_tags = {} {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.812575] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] profiler_jaeger.service_name_prefix = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.812724] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] profiler_otlp.service_name_prefix = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.812884] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.813041] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.813197] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.813331] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.813474] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.813618] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.813765] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.813909] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.814067] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.814226] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.814373] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.814545] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.814711] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.814867] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.815026] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.kombu_reconnect_splay = 0.0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.815188] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.815336] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.815482] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.815637] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.815792] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.815941] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.816102] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.816250] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.816393] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.816540] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.816720] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.816874] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.817033] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.817185] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.817334] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.817484] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.ssl = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.817632] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.817789] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.817936] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.818099] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.818254] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.ssl_version = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.818406] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.818571] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.818724] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_notifications.retry = -1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.818890] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.819081] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_messaging_notifications.transport_url = **** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.819237] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.auth_section = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.819387] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.auth_type = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.819533] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.cafile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.819678] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.certfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.819827] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.collect_timing = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.819972] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.connect_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.820128] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.connect_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.820270] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.endpoint_id = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.820425] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.endpoint_interface = publicURL {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.820569] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.endpoint_override = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.820711] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.endpoint_region_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.820856] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.endpoint_service_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.821006] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.endpoint_service_type = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.821159] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.insecure = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.821302] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.keyfile = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.821447] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.max_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.821589] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.min_version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.821735] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.region_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.821872] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.retriable_status_codes = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.822021] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.service_name = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.822198] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.service_type = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.822368] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.split_loggers = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.822515] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.status_code_retries = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.822687] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.status_code_retry_delay = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.822799] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.timeout = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.822944] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.valid_interfaces = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.823102] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_limit.version = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.823257] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_reports.file_event_handler = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.823408] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.823555] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] oslo_reports.log_dir = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.823711] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.823859] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.824008] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.824162] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plug_linux_bridge_privileged.log_daemon_traceback = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.824313] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.824459] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.824629] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.824789] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.824935] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plug_ovs_privileged.group = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.825095] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.825245] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plug_ovs_privileged.log_daemon_traceback = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.825394] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.825539] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.825683] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] vif_plug_ovs_privileged.user = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.825841] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_linux_bridge.flat_interface = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.826009] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.826175] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.826333] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.826489] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.826675] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.826837] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.826986] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.827165] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.827325] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_ovs.isolate_vif = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.827479] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.827632] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.827788] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.827943] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_ovs.ovsdb_interface = native {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.828104] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] os_vif_ovs.per_port_bridge = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.828264] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] privsep_osbrick.capabilities = [21, 2] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.828408] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] privsep_osbrick.group = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.828553] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] privsep_osbrick.helper_command = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.828713] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] privsep_osbrick.log_daemon_traceback = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.828949] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.829198] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.829368] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] privsep_osbrick.user = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.829532] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.829679] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] nova_sys_admin.group = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.829827] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] nova_sys_admin.helper_command = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.829976] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] nova_sys_admin.log_daemon_traceback = False {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.830146] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.830298] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.830442] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] nova_sys_admin.user = None {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 552.830566] env[65522]: DEBUG oslo_service.backend._eventlet.service [None req-746c87b0-4e22-49dd-ae97-a7ebca9b3b70 None None] ******************************************************************************** {{(pid=65522) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2828}} [ 552.830990] env[65522]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 553.335400] env[65522]: WARNING nova.virt.vmwareapi.driver [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 553.336110] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Getting list of instances from cluster (obj){ [ 553.336110] env[65522]: value = "domain-c8" [ 553.336110] env[65522]: _type = "ClusterComputeResource" [ 553.336110] env[65522]: } {{(pid=65522) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 553.337334] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ac5832-a4e0-48c8-81e5-24fff8ae5e2c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.346401] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Got total of 0 instances {{(pid=65522) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 553.346931] env[65522]: INFO nova.virt.node [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Generated node identity 9a0d1ac8-ed26-4657-8899-fbceab29db0d [ 553.347195] env[65522]: INFO nova.virt.node [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Wrote node identity 9a0d1ac8-ed26-4657-8899-fbceab29db0d to /opt/stack/data/n-cpu-1/compute_id [ 553.850283] env[65522]: WARNING nova.compute.manager [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Compute nodes ['9a0d1ac8-ed26-4657-8899-fbceab29db0d'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 554.856116] env[65522]: INFO nova.compute.manager [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 555.861079] env[65522]: WARNING nova.compute.manager [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 555.861079] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 555.861079] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 555.861561] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 555.861561] env[65522]: DEBUG nova.compute.resource_tracker [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 555.862188] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c92bb5-0460-431a-b0c3-a89af9eef61a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.872079] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0dfa07-b9e5-428c-acc4-9e4a857fadb9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.888380] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d12cbf1-1a22-4975-9795-c42000f671ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.895778] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12bc061a-a328-452d-a1b6-94cc0d82774c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.925024] env[65522]: DEBUG nova.compute.resource_tracker [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180527MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 555.925024] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 555.926369] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 556.429802] env[65522]: WARNING nova.compute.resource_tracker [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] No compute node record for cpu-1:9a0d1ac8-ed26-4657-8899-fbceab29db0d: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 9a0d1ac8-ed26-4657-8899-fbceab29db0d could not be found. [ 556.932872] env[65522]: INFO nova.compute.resource_tracker [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 9a0d1ac8-ed26-4657-8899-fbceab29db0d [ 558.443022] env[65522]: DEBUG nova.compute.resource_tracker [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 558.443022] env[65522]: DEBUG nova.compute.resource_tracker [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=100GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] stats={'failed_builds': '0'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 558.608082] env[65522]: INFO nova.scheduler.client.report [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] [req-1b83853e-9037-457b-867e-12bf4f276c3a] Created resource provider record via placement API for resource provider with UUID 9a0d1ac8-ed26-4657-8899-fbceab29db0d and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 558.629015] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e2ec792-742d-459e-b10e-b04047573f2e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.636560] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24da34f1-6ef9-41ca-b8a3-51920967655a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.685401] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a956890-e24e-4854-ad0d-e47956ea95f9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.696461] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad0d42c-2a92-486c-9a60-0f9f9232daf1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.714475] env[65522]: DEBUG nova.compute.provider_tree [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 559.256159] env[65522]: DEBUG nova.scheduler.client.report [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Updated inventory for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 559.256159] env[65522]: DEBUG nova.compute.provider_tree [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Updating resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d generation from 0 to 1 during operation: update_inventory {{(pid=65522) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 559.256159] env[65522]: DEBUG nova.compute.provider_tree [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 559.316019] env[65522]: DEBUG nova.compute.provider_tree [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Updating resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d generation from 1 to 2 during operation: update_traits {{(pid=65522) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 559.821023] env[65522]: DEBUG nova.compute.resource_tracker [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 559.821023] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.893s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 559.821023] env[65522]: DEBUG nova.service [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Creating RPC server for service compute {{(pid=65522) start /opt/stack/nova/nova/service.py:177}} [ 559.832791] env[65522]: DEBUG nova.service [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] Join ServiceGroup membership for this service compute {{(pid=65522) start /opt/stack/nova/nova/service.py:194}} [ 559.833182] env[65522]: DEBUG nova.servicegroup.drivers.db [None req-a7afcd73-7e9b-4d68-a644-62281a3ea970 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=65522) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 567.835761] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._sync_power_states {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 568.338870] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Getting list of instances from cluster (obj){ [ 568.338870] env[65522]: value = "domain-c8" [ 568.338870] env[65522]: _type = "ClusterComputeResource" [ 568.338870] env[65522]: } {{(pid=65522) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 568.340085] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a549ec84-7c8b-4f53-ae95-aeefbaeb9b2c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.349375] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Got total of 0 instances {{(pid=65522) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 568.349611] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 568.349920] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Getting list of instances from cluster (obj){ [ 568.349920] env[65522]: value = "domain-c8" [ 568.349920] env[65522]: _type = "ClusterComputeResource" [ 568.349920] env[65522]: } {{(pid=65522) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 568.350911] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37bf4bd4-ed59-49d7-b6c6-f0d1c0a53179 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.359628] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Got total of 0 instances {{(pid=65522) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 608.159791] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.160185] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.160324] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.160508] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.160720] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.160873] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.161454] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.161454] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 608.161454] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 608.665245] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 608.665556] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 608.665725] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 608.665852] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 608.666799] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7226ba8e-c582-415d-9e7b-070928b219e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.677517] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8f2684-1a59-46bc-8b75-972084af9615 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.692194] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d4773b-72c6-41da-b517-cc7f1e529e6c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.704741] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee15955-9b60-408f-bfe5-6e41afc8c579 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.747282] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180530MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 608.747579] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 608.747892] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 609.377769] env[65522]: INFO nova.utils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] The default thread pool MainProcess.default is initialized [ 609.379193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "57d32fb8-885e-44ba-967b-69f4bd1f5744" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 609.379193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "57d32fb8-885e-44ba-967b-69f4bd1f5744" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 609.846017] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Acquiring lock "dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 609.846017] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Lock "dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 609.884819] env[65522]: DEBUG nova.compute.manager [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 610.288351] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 57d32fb8-885e-44ba-967b-69f4bd1f5744 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 610.349405] env[65522]: DEBUG nova.compute.manager [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 610.424598] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 610.747551] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "c93b52c3-1143-44d7-8c10-ac58bc25f541" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 610.747551] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "c93b52c3-1143-44d7-8c10-ac58bc25f541" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 610.791936] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 610.792165] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 610.792622] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=100GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] stats={'failed_builds': '0'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 610.865388] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f773fa-6995-43a1-b97e-89ff69f619f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.876895] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 610.878068] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d5e086-e37c-4d5f-9354-50bfea5e57be {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.914911] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef82a4c-5c94-4fce-a814-7311cfc9a938 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.924425] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427a9c0b-fe9a-4ce2-aca4-fe337bb3123d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.941551] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.250389] env[65522]: DEBUG nova.compute.manager [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 611.445350] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 611.777338] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 611.951867] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 611.952104] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.204s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 611.952362] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.528s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 611.954392] env[65522]: INFO nova.compute.claims [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 613.029718] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Acquiring lock "fbb035b5-4562-44e3-9b95-8cf85299bb43" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.029718] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Lock "fbb035b5-4562-44e3-9b95-8cf85299bb43" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 613.101950] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6308e55-03f6-4b57-a9e6-215157b6cebd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.114211] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506654dd-f2c7-4b75-b155-ac1c64ed720e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.151387] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf96924c-fb77-452d-93dc-25ec3cd82471 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.162805] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f381a597-4ba4-41b2-bfe5-8e33142aae9f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.183851] env[65522]: DEBUG nova.compute.provider_tree [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.531869] env[65522]: DEBUG nova.compute.manager [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 613.689583] env[65522]: DEBUG nova.scheduler.client.report [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 614.070687] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 614.197911] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.245s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 614.198056] env[65522]: DEBUG nova.compute.manager [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 614.204667] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.329s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 614.206090] env[65522]: INFO nova.compute.claims [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.712545] env[65522]: DEBUG nova.compute.utils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 614.713905] env[65522]: DEBUG nova.compute.manager [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 614.715529] env[65522]: DEBUG nova.network.neutron [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 614.715529] env[65522]: WARNING neutronclient.v2_0.client [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 614.717876] env[65522]: WARNING neutronclient.v2_0.client [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 614.719777] env[65522]: WARNING openstack [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 614.720487] env[65522]: WARNING openstack [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.227811] env[65522]: DEBUG nova.compute.manager [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 615.343257] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b50801-93d7-4886-ae49-c6a5517b4991 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.352595] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0428e75-158e-496c-adf6-4e4d8deeb53d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.403062] env[65522]: DEBUG nova.policy [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '226b224888c04908a8ac4e712efb5450', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4fdbd0ed78c41a591ccc0a587313ec8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 615.405877] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fcafee8-b92c-4d27-8373-980b7a840fd9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.417684] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34a01c5-f4fb-4f58-bf7d-8b062faa73ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.432670] env[65522]: DEBUG nova.compute.provider_tree [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.941296] env[65522]: DEBUG nova.scheduler.client.report [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 616.122279] env[65522]: WARNING openstack [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.123054] env[65522]: WARNING openstack [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.262357] env[65522]: DEBUG nova.network.neutron [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Successfully created port: 7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 616.266037] env[65522]: DEBUG nova.compute.manager [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 616.303105] env[65522]: DEBUG nova.virt.hardware [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 616.303251] env[65522]: DEBUG nova.virt.hardware [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 616.303405] env[65522]: DEBUG nova.virt.hardware [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 616.303576] env[65522]: DEBUG nova.virt.hardware [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 616.303715] env[65522]: DEBUG nova.virt.hardware [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 616.305146] env[65522]: DEBUG nova.virt.hardware [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 616.305146] env[65522]: DEBUG nova.virt.hardware [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.305146] env[65522]: DEBUG nova.virt.hardware [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 616.305146] env[65522]: DEBUG nova.virt.hardware [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 616.305146] env[65522]: DEBUG nova.virt.hardware [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 616.306335] env[65522]: DEBUG nova.virt.hardware [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 616.306773] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7aa3c8a-efe0-4549-89af-379a2cb70c65 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.318546] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d615e5-0db7-4e60-b28e-4e05764ba89e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.342451] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf60800-9390-4c6e-b2c9-31f10d435c3f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.446177] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.242s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 616.447274] env[65522]: DEBUG nova.compute.manager [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 616.450366] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.673s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 616.451827] env[65522]: INFO nova.compute.claims [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 616.957596] env[65522]: DEBUG nova.compute.utils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 616.961815] env[65522]: DEBUG nova.compute.manager [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 616.961815] env[65522]: DEBUG nova.network.neutron [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 616.961941] env[65522]: WARNING neutronclient.v2_0.client [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 616.962237] env[65522]: WARNING neutronclient.v2_0.client [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 616.963187] env[65522]: WARNING openstack [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.963513] env[65522]: WARNING openstack [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 617.178639] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 617.178841] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.460706] env[65522]: DEBUG nova.policy [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3c5e099ea3446eea02b8d3cf0e232ff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b32cd6b122345db951d96d03199233c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 617.463039] env[65522]: DEBUG nova.compute.manager [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 617.668794] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2133fe-23cf-4317-b3c1-37ca4ea46bd0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.678862] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4624787d-67f2-44fc-a01b-7757dd7b35ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.682791] env[65522]: DEBUG nova.compute.manager [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 617.724283] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9673bcfc-1647-4cf8-894f-7529d343963d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.735833] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fedb6659-f97d-44f1-af17-7e5c16de11dc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.751816] env[65522]: DEBUG nova.compute.provider_tree [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.994358] env[65522]: DEBUG nova.network.neutron [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Successfully created port: e6e22697-3ece-4047-b378-b47c9c5de778 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 618.220899] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 618.228910] env[65522]: DEBUG nova.network.neutron [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Successfully updated port: 7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 618.256608] env[65522]: DEBUG nova.scheduler.client.report [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 618.473038] env[65522]: DEBUG nova.compute.manager [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 618.497990] env[65522]: DEBUG nova.virt.hardware [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 618.498373] env[65522]: DEBUG nova.virt.hardware [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 618.498541] env[65522]: DEBUG nova.virt.hardware [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 618.498740] env[65522]: DEBUG nova.virt.hardware [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 618.498914] env[65522]: DEBUG nova.virt.hardware [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 618.499077] env[65522]: DEBUG nova.virt.hardware [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 618.499308] env[65522]: DEBUG nova.virt.hardware [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.499470] env[65522]: DEBUG nova.virt.hardware [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 618.499656] env[65522]: DEBUG nova.virt.hardware [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 618.499831] env[65522]: DEBUG nova.virt.hardware [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 618.500265] env[65522]: DEBUG nova.virt.hardware [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 618.501038] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a3aa13-c082-4429-bc36-4f47c7621a47 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.510392] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f1e386-963c-4d9c-8c44-a52f8209552d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.731300] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.731408] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquired lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 618.731650] env[65522]: DEBUG nova.network.neutron [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 618.768911] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.318s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 618.769386] env[65522]: DEBUG nova.compute.manager [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 618.771968] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.704s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 618.773371] env[65522]: INFO nova.compute.claims [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.234974] env[65522]: WARNING openstack [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 619.235485] env[65522]: WARNING openstack [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 619.278493] env[65522]: DEBUG nova.compute.utils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 619.284925] env[65522]: DEBUG nova.compute.manager [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 619.285079] env[65522]: DEBUG nova.network.neutron [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 619.286048] env[65522]: WARNING neutronclient.v2_0.client [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 619.286048] env[65522]: WARNING neutronclient.v2_0.client [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 619.286197] env[65522]: WARNING openstack [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 619.286509] env[65522]: WARNING openstack [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 619.364040] env[65522]: DEBUG nova.network.neutron [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 619.438799] env[65522]: DEBUG nova.policy [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e74c132f591d4c6fbe65a75c1e1e4a24', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ecccb656b0d4c96b40b200cdcddbad5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 619.481578] env[65522]: WARNING openstack [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 619.482226] env[65522]: WARNING openstack [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 619.791209] env[65522]: DEBUG nova.compute.manager [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 619.879480] env[65522]: DEBUG nova.network.neutron [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Successfully updated port: e6e22697-3ece-4047-b378-b47c9c5de778 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 619.897176] env[65522]: DEBUG nova.network.neutron [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Successfully created port: 573a1035-e00a-4cd2-a067-934487b65b60 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 619.941424] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628b070f-46b0-40a8-ad14-910f57542850 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.952771] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63e03ea-9dd2-49be-a794-c755164ba0d8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.985634] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c24857d-0f94-4af6-a6ce-e681e19464de {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.995157] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8235bfa3-54b6-4fc8-b0b8-40c5d2b7e62e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.010098] env[65522]: DEBUG nova.compute.provider_tree [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.130699] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquiring lock "6abf58ca-ad6b-4fe8-844a-ca33e62a844d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 620.130699] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Lock "6abf58ca-ad6b-4fe8-844a-ca33e62a844d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 620.183373] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Acquiring lock "1d3a9ada-df46-46af-a7f1-a84c251af4ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 620.183617] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lock "1d3a9ada-df46-46af-a7f1-a84c251af4ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 620.367993] env[65522]: WARNING neutronclient.v2_0.client [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 620.367993] env[65522]: WARNING openstack [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 620.367993] env[65522]: WARNING openstack [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 620.386263] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Acquiring lock "refresh_cache-dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.386263] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Acquired lock "refresh_cache-dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 620.386263] env[65522]: DEBUG nova.network.neutron [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 620.513991] env[65522]: DEBUG nova.scheduler.client.report [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 620.633492] env[65522]: DEBUG nova.compute.manager [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 620.686754] env[65522]: DEBUG nova.compute.manager [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 620.804013] env[65522]: DEBUG nova.compute.manager [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 620.831233] env[65522]: DEBUG nova.virt.hardware [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 620.831608] env[65522]: DEBUG nova.virt.hardware [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 620.831829] env[65522]: DEBUG nova.virt.hardware [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 620.832131] env[65522]: DEBUG nova.virt.hardware [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 620.832336] env[65522]: DEBUG nova.virt.hardware [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 620.832545] env[65522]: DEBUG nova.virt.hardware [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 620.832805] env[65522]: DEBUG nova.virt.hardware [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 620.833018] env[65522]: DEBUG nova.virt.hardware [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 620.833244] env[65522]: DEBUG nova.virt.hardware [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 620.833461] env[65522]: DEBUG nova.virt.hardware [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 620.833692] env[65522]: DEBUG nova.virt.hardware [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 620.834721] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a5a5d7-0308-45ac-a678-9f7fe9fb4e5d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.845130] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58c431e-899e-438a-a09e-40570129f8cc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.888934] env[65522]: WARNING openstack [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 620.889385] env[65522]: WARNING openstack [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.020803] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.249s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 621.021388] env[65522]: DEBUG nova.compute.manager [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 621.024151] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.806s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 621.026198] env[65522]: INFO nova.compute.claims [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.157137] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 621.209248] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 621.494683] env[65522]: DEBUG nova.network.neutron [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Successfully updated port: 573a1035-e00a-4cd2-a067-934487b65b60 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 621.536246] env[65522]: DEBUG nova.compute.utils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 621.544368] env[65522]: DEBUG nova.network.neutron [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Updating instance_info_cache with network_info: [{"id": "7d49f270-6cbd-4767-91fc-f899c46564c9", "address": "fa:16:3e:af:04:9f", "network": {"id": "6a60138b-7f03-4817-ac9a-1facdc1cc0cc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1344754077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4fdbd0ed78c41a591ccc0a587313ec8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f85835c8-5d0c-4b2f-97c4-6c4006580f79", "external-id": "nsx-vlan-transportzone-245", "segmentation_id": 245, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d49f270-6c", "ovs_interfaceid": "7d49f270-6cbd-4767-91fc-f899c46564c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 621.544868] env[65522]: DEBUG nova.compute.manager [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 621.545469] env[65522]: DEBUG nova.network.neutron [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 621.545900] env[65522]: WARNING neutronclient.v2_0.client [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 621.546506] env[65522]: WARNING neutronclient.v2_0.client [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 621.547111] env[65522]: WARNING openstack [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.547605] env[65522]: WARNING openstack [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.599650] env[65522]: DEBUG nova.network.neutron [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 621.708686] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-708c8022-0bab-47ab-b55c-044d6ba1b414 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.721201] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a86743-9a09-436f-8354-4e4ea0e4fc05 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.754335] env[65522]: WARNING openstack [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.754728] env[65522]: WARNING openstack [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.764126] env[65522]: DEBUG nova.policy [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4def821de0414e84977d8c6b5a3c13d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f453b4359a94e54aca88fa2b9459476', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 621.766334] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a5a3693-fb03-4481-a893-4fa8bf11f92a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.777374] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64c1812-4b0b-4007-a30d-74f92023c7ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.791056] env[65522]: DEBUG nova.compute.provider_tree [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.997221] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "refresh_cache-c93b52c3-1143-44d7-8c10-ac58bc25f541" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.997221] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "refresh_cache-c93b52c3-1143-44d7-8c10-ac58bc25f541" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 621.997517] env[65522]: DEBUG nova.network.neutron [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 622.041906] env[65522]: DEBUG nova.compute.manager [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 622.061662] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Releasing lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 622.062382] env[65522]: DEBUG nova.compute.manager [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Instance network_info: |[{"id": "7d49f270-6cbd-4767-91fc-f899c46564c9", "address": "fa:16:3e:af:04:9f", "network": {"id": "6a60138b-7f03-4817-ac9a-1facdc1cc0cc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1344754077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4fdbd0ed78c41a591ccc0a587313ec8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f85835c8-5d0c-4b2f-97c4-6c4006580f79", "external-id": "nsx-vlan-transportzone-245", "segmentation_id": 245, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d49f270-6c", "ovs_interfaceid": "7d49f270-6cbd-4767-91fc-f899c46564c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 622.062690] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:af:04:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f85835c8-5d0c-4b2f-97c4-6c4006580f79', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d49f270-6cbd-4767-91fc-f899c46564c9', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 622.076463] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 622.076740] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-084b70b9-563b-4479-8056-72deeb4473c4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.090588] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Created folder: OpenStack in parent group-v4. [ 622.090918] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Creating folder: Project (e4fdbd0ed78c41a591ccc0a587313ec8). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 622.091121] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8befbefa-e51c-42bf-a573-14ae1a84f859 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.105866] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Created folder: Project (e4fdbd0ed78c41a591ccc0a587313ec8) in parent group-v994660. [ 622.106062] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Creating folder: Instances. Parent ref: group-v994661. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 622.106334] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4494f3db-b748-4922-8e89-791dffa161ad {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.117558] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Created folder: Instances in parent group-v994661. [ 622.117861] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 622.118113] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 622.118367] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3866930f-5a40-4cb6-890d-bcfac370a8ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.140452] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 622.140452] env[65522]: value = "task-5113479" [ 622.140452] env[65522]: _type = "Task" [ 622.140452] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.150191] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113479, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.294711] env[65522]: DEBUG nova.scheduler.client.report [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 622.443135] env[65522]: DEBUG nova.network.neutron [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Successfully created port: 43c29c73-562e-48ba-886e-8e42d9c04074 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 622.472568] env[65522]: WARNING neutronclient.v2_0.client [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 622.472568] env[65522]: WARNING openstack [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.472568] env[65522]: WARNING openstack [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.501174] env[65522]: WARNING openstack [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.501560] env[65522]: WARNING openstack [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.599403] env[65522]: DEBUG nova.network.neutron [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 622.651617] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113479, 'name': CreateVM_Task, 'duration_secs': 0.455126} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.651836] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 622.652298] env[65522]: WARNING neutronclient.v2_0.client [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 622.652937] env[65522]: WARNING openstack [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.653229] env[65522]: WARNING openstack [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.697174] env[65522]: DEBUG nova.network.neutron [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Updating instance_info_cache with network_info: [{"id": "e6e22697-3ece-4047-b378-b47c9c5de778", "address": "fa:16:3e:1a:4e:a7", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.26", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6e22697-3e", "ovs_interfaceid": "e6e22697-3ece-4047-b378-b47c9c5de778", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 622.719252] env[65522]: WARNING openstack [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.720090] env[65522]: WARNING openstack [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.796271] env[65522]: DEBUG oslo_vmware.service [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bf4c715-24fd-4d49-929e-3fda5625babd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.799974] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.776s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 622.800419] env[65522]: DEBUG nova.compute.manager [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 622.807282] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.647s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 622.809040] env[65522]: INFO nova.compute.claims [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.814101] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.814180] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 622.814834] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 622.817532] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3643c30e-6511-4d7d-a3e5-750d000ab7e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.827851] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 622.827851] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b4d407-1bf3-d1d6-08d7-24f697843866" [ 622.827851] env[65522]: _type = "Task" [ 622.827851] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.840586] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b4d407-1bf3-d1d6-08d7-24f697843866, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.860755] env[65522]: WARNING neutronclient.v2_0.client [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 622.861540] env[65522]: WARNING openstack [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.861930] env[65522]: WARNING openstack [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 623.027087] env[65522]: DEBUG nova.network.neutron [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Updating instance_info_cache with network_info: [{"id": "573a1035-e00a-4cd2-a067-934487b65b60", "address": "fa:16:3e:8a:04:eb", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap573a1035-e0", "ovs_interfaceid": "573a1035-e00a-4cd2-a067-934487b65b60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 623.051503] env[65522]: DEBUG nova.compute.manager [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 623.080564] env[65522]: DEBUG nova.virt.hardware [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 623.080844] env[65522]: DEBUG nova.virt.hardware [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 623.081038] env[65522]: DEBUG nova.virt.hardware [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 623.081231] env[65522]: DEBUG nova.virt.hardware [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 623.081373] env[65522]: DEBUG nova.virt.hardware [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 623.081513] env[65522]: DEBUG nova.virt.hardware [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 623.081718] env[65522]: DEBUG nova.virt.hardware [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 623.081877] env[65522]: DEBUG nova.virt.hardware [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 623.082058] env[65522]: DEBUG nova.virt.hardware [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 623.082219] env[65522]: DEBUG nova.virt.hardware [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 623.082423] env[65522]: DEBUG nova.virt.hardware [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 623.083648] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5885b782-f9c3-433e-bcb9-2b83f98ce447 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.092897] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c82e83b1-cd1d-40f3-9bff-f091bb65f209 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.199874] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Releasing lock "refresh_cache-dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 623.201025] env[65522]: DEBUG nova.compute.manager [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Instance network_info: |[{"id": "e6e22697-3ece-4047-b378-b47c9c5de778", "address": "fa:16:3e:1a:4e:a7", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.26", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6e22697-3e", "ovs_interfaceid": "e6e22697-3ece-4047-b378-b47c9c5de778", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 623.201557] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:4e:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e6e22697-3ece-4047-b378-b47c9c5de778', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 623.209151] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Creating folder: Project (3b32cd6b122345db951d96d03199233c). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 623.209473] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-244bd479-2938-477a-9705-90cad04074c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.221899] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Created folder: Project (3b32cd6b122345db951d96d03199233c) in parent group-v994660. [ 623.221899] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Creating folder: Instances. Parent ref: group-v994664. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 623.221899] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1d0c2d91-63b8-4d43-baac-ad52311ee935 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.231595] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Created folder: Instances in parent group-v994664. [ 623.231887] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 623.232070] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 623.233027] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b0440ef5-ba44-497a-a0fa-7eefaa77ecbf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.253739] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 623.253739] env[65522]: value = "task-5113482" [ 623.253739] env[65522]: _type = "Task" [ 623.253739] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.263850] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113482, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.321662] env[65522]: DEBUG nova.compute.utils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 623.323670] env[65522]: DEBUG nova.compute.manager [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 623.324122] env[65522]: DEBUG nova.network.neutron [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 623.324437] env[65522]: WARNING neutronclient.v2_0.client [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 623.324752] env[65522]: WARNING neutronclient.v2_0.client [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 623.325436] env[65522]: WARNING openstack [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 623.325862] env[65522]: WARNING openstack [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 623.347222] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquiring lock "bc40780e-43e2-4058-ab6f-d0d65e839128" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 623.347466] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lock "bc40780e-43e2-4058-ab6f-d0d65e839128" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 623.348982] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 623.349185] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 623.349404] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.349543] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 623.349960] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 623.350756] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-239c056c-e221-4820-909b-44d2eaae9fbe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.374076] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 623.374340] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 623.375326] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924662d0-9621-4798-8c86-f8e31836a56f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.384265] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a61a885c-2fc8-4387-844e-d447ba2b3cd2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.390664] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 623.390664] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520c3329-92af-fd43-c52a-b23d18ec52a9" [ 623.390664] env[65522]: _type = "Task" [ 623.390664] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.400524] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520c3329-92af-fd43-c52a-b23d18ec52a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.416979] env[65522]: DEBUG nova.policy [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b491939798e4481fb433ffb81f366b25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '068c2387de8c406194d9b1762c7292a7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 623.504531] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Acquiring lock "a09c12be-0aaa-4934-a1b3-5af29feba8a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 623.504531] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Lock "a09c12be-0aaa-4934-a1b3-5af29feba8a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 623.530673] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "refresh_cache-c93b52c3-1143-44d7-8c10-ac58bc25f541" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 623.531163] env[65522]: DEBUG nova.compute.manager [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Instance network_info: |[{"id": "573a1035-e00a-4cd2-a067-934487b65b60", "address": "fa:16:3e:8a:04:eb", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap573a1035-e0", "ovs_interfaceid": "573a1035-e00a-4cd2-a067-934487b65b60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 623.532056] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:04:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6fb0104-186b-4288-b87e-634893f46f01', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '573a1035-e00a-4cd2-a067-934487b65b60', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 623.541230] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Creating folder: Project (6ecccb656b0d4c96b40b200cdcddbad5). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 623.542225] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff182b45-5e52-4155-8928-a4c33b028c7c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.552647] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Created folder: Project (6ecccb656b0d4c96b40b200cdcddbad5) in parent group-v994660. [ 623.552849] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Creating folder: Instances. Parent ref: group-v994667. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 623.553131] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-049c17bd-b729-4f31-8d5b-b24bd987ecec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.571958] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Created folder: Instances in parent group-v994667. [ 623.572293] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 623.572594] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 623.572742] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-77ffe7cd-3253-4f9c-85f3-e7984abc745c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.595740] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 623.595740] env[65522]: value = "task-5113485" [ 623.595740] env[65522]: _type = "Task" [ 623.595740] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.604714] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113485, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.767102] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113482, 'name': CreateVM_Task, 'duration_secs': 0.376813} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.767296] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 623.767960] env[65522]: WARNING neutronclient.v2_0.client [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 623.768361] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.768525] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 623.768865] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 623.769138] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-398c51c6-f7af-4e47-bf73-116108641640 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.775045] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Waiting for the task: (returnval){ [ 623.775045] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522bf9e0-04d7-015a-8984-b9ac927395be" [ 623.775045] env[65522]: _type = "Task" [ 623.775045] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.786130] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522bf9e0-04d7-015a-8984-b9ac927395be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.834932] env[65522]: DEBUG nova.compute.manager [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 623.853025] env[65522]: DEBUG nova.compute.manager [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 623.905253] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Preparing fetch location {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 623.905580] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Creating directory with path [datastore1] vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5/a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 623.905883] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8bb666ef-e32d-4e9b-9ccc-75bcd4b81612 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.911783] env[65522]: DEBUG nova.network.neutron [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Successfully created port: d24163e8-4f08-4908-9307-c4edb0d4ffb7 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 623.927375] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Created directory with path [datastore1] vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5/a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 623.927375] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Fetch image to [datastore1] vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 623.927643] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Downloading image file data a889db67-7337-4e32-8e34-642f34402926 to [datastore1] vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk on the data store datastore1 {{(pid=65522) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 623.928336] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282c970c-5de1-42a5-b3c6-9c88cb63edd4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.936285] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22d4b90-6b58-453a-80fa-cef7c37c12ce {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.951933] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454b5e29-d0eb-43f8-b158-399ee65d0408 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.996094] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d558def-a1d3-4023-b619-6de8aed71b9d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.005026] env[65522]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-5929d4bc-fba4-471f-8be9-3442924425a6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.008408] env[65522]: DEBUG nova.compute.manager [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 624.029849] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe86a59-0766-4076-a058-f4c95a16dc43 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.048181] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04618fd-a8b0-47a8-a98e-44aafcfc57f4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.091280] env[65522]: DEBUG nova.network.neutron [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Successfully updated port: 43c29c73-562e-48ba-886e-8e42d9c04074 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 624.093949] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c5e547-674c-435c-b456-e85a4251fd28 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.111354] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc529c7a-8c32-4663-969d-d774115de64f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.116411] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Downloading image file data a889db67-7337-4e32-8e34-642f34402926 to the data store datastore1 {{(pid=65522) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 624.124507] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113485, 'name': CreateVM_Task, 'duration_secs': 0.35899} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.133109] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 624.133654] env[65522]: DEBUG nova.compute.provider_tree [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.135151] env[65522]: WARNING neutronclient.v2_0.client [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 624.135507] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.204301] env[65522]: DEBUG oslo_vmware.rw_handles [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=65522) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 624.210533] env[65522]: DEBUG nova.compute.manager [req-74df9d33-8cfa-415c-9a2c-c80b81175dc0 req-48114413-e30e-4dc3-bd03-924292578bd2 service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Received event network-vif-plugged-7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 624.210533] env[65522]: DEBUG oslo_concurrency.lockutils [req-74df9d33-8cfa-415c-9a2c-c80b81175dc0 req-48114413-e30e-4dc3-bd03-924292578bd2 service nova] Acquiring lock "57d32fb8-885e-44ba-967b-69f4bd1f5744-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 624.210533] env[65522]: DEBUG oslo_concurrency.lockutils [req-74df9d33-8cfa-415c-9a2c-c80b81175dc0 req-48114413-e30e-4dc3-bd03-924292578bd2 service nova] Lock "57d32fb8-885e-44ba-967b-69f4bd1f5744-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 624.210533] env[65522]: DEBUG oslo_concurrency.lockutils [req-74df9d33-8cfa-415c-9a2c-c80b81175dc0 req-48114413-e30e-4dc3-bd03-924292578bd2 service nova] Lock "57d32fb8-885e-44ba-967b-69f4bd1f5744-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 624.210533] env[65522]: DEBUG nova.compute.manager [req-74df9d33-8cfa-415c-9a2c-c80b81175dc0 req-48114413-e30e-4dc3-bd03-924292578bd2 service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] No waiting events found dispatching network-vif-plugged-7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 624.211130] env[65522]: WARNING nova.compute.manager [req-74df9d33-8cfa-415c-9a2c-c80b81175dc0 req-48114413-e30e-4dc3-bd03-924292578bd2 service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Received unexpected event network-vif-plugged-7d49f270-6cbd-4767-91fc-f899c46564c9 for instance with vm_state building and task_state spawning. [ 624.292050] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 624.292050] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 624.292050] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.292856] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.292974] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 624.293281] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c00411a-489e-4cb2-b1ac-bcb27aaf95d2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.299740] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 624.299740] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521700c1-3e47-3b33-9feb-0172eee0121f" [ 624.299740] env[65522]: _type = "Task" [ 624.299740] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.312212] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521700c1-3e47-3b33-9feb-0172eee0121f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.381402] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 624.532446] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 624.598429] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Acquiring lock "refresh_cache-fbb035b5-4562-44e3-9b95-8cf85299bb43" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.598429] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Acquired lock "refresh_cache-fbb035b5-4562-44e3-9b95-8cf85299bb43" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.598683] env[65522]: DEBUG nova.network.neutron [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 624.637648] env[65522]: DEBUG nova.scheduler.client.report [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 624.714621] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "7c6f8218-602d-44f3-8012-de5a96972785" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 624.714891] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "7c6f8218-602d-44f3-8012-de5a96972785" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 624.821762] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 624.822065] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 624.822286] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.849565] env[65522]: DEBUG nova.compute.manager [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 624.889362] env[65522]: DEBUG nova.virt.hardware [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 624.889582] env[65522]: DEBUG nova.virt.hardware [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 624.889582] env[65522]: DEBUG nova.virt.hardware [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 624.889736] env[65522]: DEBUG nova.virt.hardware [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 624.889878] env[65522]: DEBUG nova.virt.hardware [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 624.890072] env[65522]: DEBUG nova.virt.hardware [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 624.890782] env[65522]: DEBUG nova.virt.hardware [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 624.890782] env[65522]: DEBUG nova.virt.hardware [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 624.890782] env[65522]: DEBUG nova.virt.hardware [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 624.890782] env[65522]: DEBUG nova.virt.hardware [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 624.890990] env[65522]: DEBUG nova.virt.hardware [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 624.891905] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9759d23d-902b-425a-9f09-94ae983b993a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.905872] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703e8784-8a48-4187-a895-883045a0d823 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.930279] env[65522]: DEBUG oslo_vmware.rw_handles [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Completed reading data from the image iterator. {{(pid=65522) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 624.930489] env[65522]: DEBUG oslo_vmware.rw_handles [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 625.096949] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Downloaded image file data a889db67-7337-4e32-8e34-642f34402926 to vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk on the data store datastore1 {{(pid=65522) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 625.098555] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Caching image {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 625.098830] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Copying Virtual Disk [datastore1] vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk to [datastore1] vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 625.099240] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da4ce1c7-a34f-4d5e-b29c-1bf951e6d3b9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.102068] env[65522]: WARNING openstack [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.102423] env[65522]: WARNING openstack [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.115632] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 625.115632] env[65522]: value = "task-5113486" [ 625.115632] env[65522]: _type = "Task" [ 625.115632] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.125610] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113486, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.144665] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.341s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 625.145228] env[65522]: DEBUG nova.compute.manager [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 625.152685] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.943s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 625.154433] env[65522]: INFO nova.compute.claims [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 625.158493] env[65522]: DEBUG nova.network.neutron [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 625.217446] env[65522]: DEBUG nova.compute.manager [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 625.234948] env[65522]: WARNING openstack [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.235425] env[65522]: WARNING openstack [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.394621] env[65522]: WARNING neutronclient.v2_0.client [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 625.395519] env[65522]: WARNING openstack [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.396877] env[65522]: WARNING openstack [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.493722] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Acquiring lock "3f749f68-879f-44f3-8477-7ecea947078c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 625.493977] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Lock "3f749f68-879f-44f3-8477-7ecea947078c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 625.631313] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113486, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.655159] env[65522]: DEBUG nova.compute.utils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 625.657484] env[65522]: DEBUG nova.compute.manager [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 625.657852] env[65522]: DEBUG nova.network.neutron [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 625.658553] env[65522]: WARNING neutronclient.v2_0.client [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 625.658956] env[65522]: WARNING neutronclient.v2_0.client [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 625.659615] env[65522]: WARNING openstack [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.660020] env[65522]: WARNING openstack [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.743015] env[65522]: DEBUG nova.network.neutron [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Successfully updated port: d24163e8-4f08-4908-9307-c4edb0d4ffb7 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 625.748699] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 625.784175] env[65522]: DEBUG nova.policy [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bec6f0cea08240dc917d24fe566358ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a47babe780ec4ae7b6c40f4531f09a38', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 625.940342] env[65522]: DEBUG nova.network.neutron [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Updating instance_info_cache with network_info: [{"id": "43c29c73-562e-48ba-886e-8e42d9c04074", "address": "fa:16:3e:7a:51:6e", "network": {"id": "2f4931f5-90c6-44a9-922e-e942d59ba127", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1285529142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f453b4359a94e54aca88fa2b9459476", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9a1e09ef-7c9c-45d9-9bf4-55b913524948", "external-id": "nsx-vlan-transportzone-466", "segmentation_id": 466, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c29c73-56", "ovs_interfaceid": "43c29c73-562e-48ba-886e-8e42d9c04074", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 626.129155] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113486, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.667182} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.129419] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Copied Virtual Disk [datastore1] vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk to [datastore1] vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 626.129596] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Deleting the datastore file [datastore1] vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 626.129863] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1bcd2e1-04b4-446a-9576-4a30396f43b4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.137589] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 626.137589] env[65522]: value = "task-5113487" [ 626.137589] env[65522]: _type = "Task" [ 626.137589] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.147686] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113487, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.172675] env[65522]: DEBUG nova.compute.manager [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 626.229367] env[65522]: DEBUG nova.network.neutron [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Successfully created port: dcf081b2-7a3e-4efd-84fa-bbf035dfd690 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 626.247288] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.247473] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 626.247746] env[65522]: DEBUG nova.network.neutron [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 626.382931] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eafc165-3871-441e-af95-1a2a5f0af15b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.392912] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7e2548-c358-4102-bf14-d94799cbb1e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.430470] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387c390f-ea48-404c-9ccc-b6e80a15c4d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.441895] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0c1c19-152a-4b23-8994-5f4e607fd738 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.448069] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Releasing lock "refresh_cache-fbb035b5-4562-44e3-9b95-8cf85299bb43" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 626.448069] env[65522]: DEBUG nova.compute.manager [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Instance network_info: |[{"id": "43c29c73-562e-48ba-886e-8e42d9c04074", "address": "fa:16:3e:7a:51:6e", "network": {"id": "2f4931f5-90c6-44a9-922e-e942d59ba127", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1285529142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f453b4359a94e54aca88fa2b9459476", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9a1e09ef-7c9c-45d9-9bf4-55b913524948", "external-id": "nsx-vlan-transportzone-466", "segmentation_id": 466, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c29c73-56", "ovs_interfaceid": "43c29c73-562e-48ba-886e-8e42d9c04074", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 626.448327] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:51:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9a1e09ef-7c9c-45d9-9bf4-55b913524948', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43c29c73-562e-48ba-886e-8e42d9c04074', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 626.456744] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Creating folder: Project (2f453b4359a94e54aca88fa2b9459476). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 626.457288] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-90ae9654-c563-4c4a-ad11-fbf5dd98a9b3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.469025] env[65522]: DEBUG nova.compute.provider_tree [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.482118] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Created folder: Project (2f453b4359a94e54aca88fa2b9459476) in parent group-v994660. [ 626.483061] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Creating folder: Instances. Parent ref: group-v994670. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 626.483061] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-64361ef8-745b-44cf-a0d7-07e4766f7e0e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.495211] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Created folder: Instances in parent group-v994670. [ 626.495550] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 626.495757] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 626.495979] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ecefea2-1b5f-4854-8e77-50a0bb73b29f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.518446] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 626.518446] env[65522]: value = "task-5113490" [ 626.518446] env[65522]: _type = "Task" [ 626.518446] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.529403] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113490, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.572205] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Acquiring lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 626.572837] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 626.654989] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113487, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024879} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.655478] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 626.655745] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Moving file from [datastore1] vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5/a889db67-7337-4e32-8e34-642f34402926 to [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926. {{(pid=65522) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 626.656064] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-397af000-9a57-4995-a279-e667bcd7618b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.664212] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 626.664212] env[65522]: value = "task-5113491" [ 626.664212] env[65522]: _type = "Task" [ 626.664212] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.675088] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113491, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.755068] env[65522]: WARNING openstack [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 626.757910] env[65522]: WARNING openstack [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 626.973055] env[65522]: DEBUG nova.scheduler.client.report [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 627.031245] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113490, 'name': CreateVM_Task, 'duration_secs': 0.357968} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.031976] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 627.032497] env[65522]: WARNING neutronclient.v2_0.client [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 627.032849] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.033037] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 627.033354] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 627.033672] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afb2725d-1da7-4936-8148-fda091d258ac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.039126] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Waiting for the task: (returnval){ [ 627.039126] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526ae24d-0eed-04b3-9b56-7e8f534ccd39" [ 627.039126] env[65522]: _type = "Task" [ 627.039126] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.049995] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526ae24d-0eed-04b3-9b56-7e8f534ccd39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.055413] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "6c8165f5-6769-4954-b9ac-de8551954377" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 627.056345] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "6c8165f5-6769-4954-b9ac-de8551954377" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 627.175623] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113491, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.032543} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.176135] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] File moved {{(pid=65522) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 627.176246] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Cleaning up location [datastore1] vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 627.176489] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Deleting the datastore file [datastore1] vmware_temp/daaa700e-4b0a-4385-9c9f-087ac003a9e5 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 627.176800] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ddd7d308-99df-462b-8a86-94d46eb5287d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.182102] env[65522]: DEBUG nova.compute.manager [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 627.186120] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 627.186120] env[65522]: value = "task-5113492" [ 627.186120] env[65522]: _type = "Task" [ 627.186120] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.197910] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113492, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.218165] env[65522]: DEBUG nova.virt.hardware [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 627.218165] env[65522]: DEBUG nova.virt.hardware [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 627.218165] env[65522]: DEBUG nova.virt.hardware [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 627.218461] env[65522]: DEBUG nova.virt.hardware [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 627.218792] env[65522]: DEBUG nova.virt.hardware [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 627.219127] env[65522]: DEBUG nova.virt.hardware [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 627.219480] env[65522]: DEBUG nova.virt.hardware [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 627.219779] env[65522]: DEBUG nova.virt.hardware [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 627.220233] env[65522]: DEBUG nova.virt.hardware [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 627.220556] env[65522]: DEBUG nova.virt.hardware [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 627.220866] env[65522]: DEBUG nova.virt.hardware [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 627.221967] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6390a78-20c5-4a80-9523-fb8c9de5003f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.230453] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec829d93-85c7-485d-bdba-fe5eef0e5582 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.470675] env[65522]: DEBUG nova.network.neutron [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 627.477669] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.325s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 627.478371] env[65522]: DEBUG nova.compute.manager [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 627.481516] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.100s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 627.483281] env[65522]: INFO nova.compute.claims [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.537776] env[65522]: WARNING openstack [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.538088] env[65522]: WARNING openstack [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 627.556246] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526ae24d-0eed-04b3-9b56-7e8f534ccd39, 'name': SearchDatastore_Task, 'duration_secs': 0.01006} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.556569] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 627.556752] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 627.557023] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.701117] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113492, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027065} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.701654] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 627.702089] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b7b0685-b229-4c0d-839a-4fd5327d9d25 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.708179] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 627.708179] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52285ceb-455d-c08a-febc-498bff15b61a" [ 627.708179] env[65522]: _type = "Task" [ 627.708179] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.718495] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52285ceb-455d-c08a-febc-498bff15b61a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.766737] env[65522]: WARNING neutronclient.v2_0.client [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 627.767408] env[65522]: WARNING openstack [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.768913] env[65522]: WARNING openstack [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 627.899806] env[65522]: DEBUG nova.network.neutron [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Updating instance_info_cache with network_info: [{"id": "d24163e8-4f08-4908-9307-c4edb0d4ffb7", "address": "fa:16:3e:e2:90:31", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd24163e8-4f", "ovs_interfaceid": "d24163e8-4f08-4908-9307-c4edb0d4ffb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 627.992209] env[65522]: DEBUG nova.compute.utils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 627.997889] env[65522]: DEBUG nova.compute.manager [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 627.998141] env[65522]: DEBUG nova.network.neutron [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 627.998569] env[65522]: WARNING neutronclient.v2_0.client [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 627.998850] env[65522]: WARNING neutronclient.v2_0.client [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 627.999451] env[65522]: WARNING openstack [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.999772] env[65522]: WARNING openstack [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.049756] env[65522]: DEBUG nova.network.neutron [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Successfully updated port: dcf081b2-7a3e-4efd-84fa-bbf035dfd690 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 628.138017] env[65522]: DEBUG nova.policy [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8123b3e1a9f419392d464a707761eb7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '345c026cba1d40c59cab5aa5b477b833', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 628.223295] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52285ceb-455d-c08a-febc-498bff15b61a, 'name': SearchDatastore_Task, 'duration_secs': 0.009897} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.223295] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 628.223295] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 57d32fb8-885e-44ba-967b-69f4bd1f5744/57d32fb8-885e-44ba-967b-69f4bd1f5744.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 628.223295] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 628.223644] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 628.223644] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c4fcac28-4f21-4f8d-b4dc-2c45169b855e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.225029] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40264848-f1c6-4a21-b87b-5234cdc9904a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.232439] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 628.232439] env[65522]: value = "task-5113493" [ 628.232439] env[65522]: _type = "Task" [ 628.232439] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.237123] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 628.237324] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 628.238705] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-630ca94f-0345-4f67-9f6f-413fdbaa0c1a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.244614] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113493, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.248477] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Waiting for the task: (returnval){ [ 628.248477] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5226bcd4-a07b-0283-fa45-d149da392074" [ 628.248477] env[65522]: _type = "Task" [ 628.248477] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.258348] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5226bcd4-a07b-0283-fa45-d149da392074, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.403606] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 628.404235] env[65522]: DEBUG nova.compute.manager [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Instance network_info: |[{"id": "d24163e8-4f08-4908-9307-c4edb0d4ffb7", "address": "fa:16:3e:e2:90:31", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd24163e8-4f", "ovs_interfaceid": "d24163e8-4f08-4908-9307-c4edb0d4ffb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 628.405142] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e2:90:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4345ef6-a7c8-4c1c-badf-a0d4f578b61c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd24163e8-4f08-4908-9307-c4edb0d4ffb7', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 628.413636] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Creating folder: Project (068c2387de8c406194d9b1762c7292a7). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 628.414267] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03fb1484-095a-4dd1-b94f-4699f90ba09d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.430028] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Created folder: Project (068c2387de8c406194d9b1762c7292a7) in parent group-v994660. [ 628.430028] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Creating folder: Instances. Parent ref: group-v994673. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 628.430028] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-41d502c3-94b6-492b-a12b-e43795c84291 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.445497] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Created folder: Instances in parent group-v994673. [ 628.445497] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 628.445497] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 628.445497] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d226a47-9ac4-4b60-806d-393ff95b6f1f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.474202] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 628.474202] env[65522]: value = "task-5113496" [ 628.474202] env[65522]: _type = "Task" [ 628.474202] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.486725] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113496, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.498721] env[65522]: DEBUG nova.compute.manager [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 628.536989] env[65522]: DEBUG nova.network.neutron [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Successfully created port: 3896379b-7eb4-4dc5-819d-f68421a44dab {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 628.554378] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquiring lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.554378] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquired lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 628.554378] env[65522]: DEBUG nova.network.neutron [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 628.753175] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113493, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.769308] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5226bcd4-a07b-0283-fa45-d149da392074, 'name': SearchDatastore_Task, 'duration_secs': 0.009487} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.770514] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8b87d83-d1db-4c84-8fef-7434eece093b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.777439] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Waiting for the task: (returnval){ [ 628.777439] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52893698-0356-65ca-0822-326ba7ea21db" [ 628.777439] env[65522]: _type = "Task" [ 628.777439] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.790101] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52893698-0356-65ca-0822-326ba7ea21db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.819375] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5535e9-3d66-4c53-8ec4-ce9bb6cc738d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.828287] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7987661e-2979-43f6-951a-2e0d290f259f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.863979] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c1655a6-0017-43f4-a3c8-74e2bb27da28 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.872806] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e550e2-57c4-4892-b958-ee395e40f289 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.889597] env[65522]: DEBUG nova.compute.provider_tree [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.985640] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113496, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.061338] env[65522]: WARNING openstack [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 629.061713] env[65522]: WARNING openstack [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 629.251183] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113493, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542884} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.251493] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 57d32fb8-885e-44ba-967b-69f4bd1f5744/57d32fb8-885e-44ba-967b-69f4bd1f5744.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 629.251720] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 629.251980] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19c12811-1501-42ef-b659-d2114146c3c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.259636] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 629.259636] env[65522]: value = "task-5113497" [ 629.259636] env[65522]: _type = "Task" [ 629.259636] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.270262] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113497, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.290505] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52893698-0356-65ca-0822-326ba7ea21db, 'name': SearchDatastore_Task, 'duration_secs': 0.011817} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.290786] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 629.291068] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7/dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 629.291365] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 629.291556] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 629.291793] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f078c4ab-5c10-42c0-8dc4-19079724d158 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.294623] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-145a5f47-522c-4391-98b4-a59479e6a153 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.303430] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Waiting for the task: (returnval){ [ 629.303430] env[65522]: value = "task-5113498" [ 629.303430] env[65522]: _type = "Task" [ 629.303430] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.308234] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 629.308448] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 629.309717] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1909458-c163-44fb-b0a4-b3384bb2f3d0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.315488] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113498, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.319771] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 629.319771] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d7bb63-8825-0bdd-dcdf-4b7e4093d453" [ 629.319771] env[65522]: _type = "Task" [ 629.319771] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.329532] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d7bb63-8825-0bdd-dcdf-4b7e4093d453, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.393621] env[65522]: DEBUG nova.scheduler.client.report [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 629.452751] env[65522]: DEBUG nova.network.neutron [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 629.489933] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113496, 'name': CreateVM_Task, 'duration_secs': 0.533599} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.490340] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 629.490646] env[65522]: WARNING neutronclient.v2_0.client [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 629.491063] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.491218] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 629.491554] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 629.491877] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95067027-0ce1-43cd-805d-73b1a7c7ef51 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.502438] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 629.502438] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e69f98-56af-8426-c45a-787a694b8aa5" [ 629.502438] env[65522]: _type = "Task" [ 629.502438] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.515607] env[65522]: DEBUG nova.compute.manager [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 629.517886] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e69f98-56af-8426-c45a-787a694b8aa5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.577172] env[65522]: DEBUG nova.virt.hardware [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 629.577468] env[65522]: DEBUG nova.virt.hardware [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 629.577691] env[65522]: DEBUG nova.virt.hardware [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 629.577878] env[65522]: DEBUG nova.virt.hardware [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 629.578096] env[65522]: DEBUG nova.virt.hardware [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 629.578279] env[65522]: DEBUG nova.virt.hardware [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 629.578495] env[65522]: DEBUG nova.virt.hardware [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.579879] env[65522]: DEBUG nova.virt.hardware [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 629.580090] env[65522]: DEBUG nova.virt.hardware [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 629.580264] env[65522]: DEBUG nova.virt.hardware [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 629.580440] env[65522]: DEBUG nova.virt.hardware [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 629.581374] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5caf01b2-0587-4347-b90f-9bb0eeb654a1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.590830] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22161b69-f5ea-451d-9b4d-21b4f9421bf2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.623673] env[65522]: WARNING openstack [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 629.624181] env[65522]: WARNING openstack [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 629.772739] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113497, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066442} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.773172] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 629.774071] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f16bb30b-1ab7-43f1-879f-45442564411e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.804364] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] 57d32fb8-885e-44ba-967b-69f4bd1f5744/57d32fb8-885e-44ba-967b-69f4bd1f5744.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 629.804715] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5cc34a37-d785-4f52-bc00-bd33427caada {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.831146] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 629.831146] env[65522]: value = "task-5113499" [ 629.831146] env[65522]: _type = "Task" [ 629.831146] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.840176] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d7bb63-8825-0bdd-dcdf-4b7e4093d453, 'name': SearchDatastore_Task, 'duration_secs': 0.032676} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.840428] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113498, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.846517] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d70311bd-3272-42f8-8a1e-6dc6272eae24 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.860505] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.862639] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 629.862639] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5228f84d-455a-6cfc-c24a-9456dab0fcaf" [ 629.862639] env[65522]: _type = "Task" [ 629.862639] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.873789] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5228f84d-455a-6cfc-c24a-9456dab0fcaf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.880172] env[65522]: WARNING neutronclient.v2_0.client [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 629.880596] env[65522]: WARNING openstack [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 629.880943] env[65522]: WARNING openstack [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 629.898912] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 629.899475] env[65522]: DEBUG nova.compute.manager [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 629.902129] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.370s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 629.903814] env[65522]: INFO nova.compute.claims [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.013833] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e69f98-56af-8426-c45a-787a694b8aa5, 'name': SearchDatastore_Task, 'duration_secs': 0.013155} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.014528] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 630.014528] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 630.014694] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.121346] env[65522]: DEBUG nova.network.neutron [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Updating instance_info_cache with network_info: [{"id": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "address": "fa:16:3e:9b:22:d2", "network": {"id": "b1e15367-b861-4a6c-8f94-921ed2b0a66f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1304672897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a47babe780ec4ae7b6c40f4531f09a38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "79c2e589-f55b-4843-8d99-2e565be16706", "external-id": "nsx-vlan-transportzone-858", "segmentation_id": 858, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcf081b2-7a", "ovs_interfaceid": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 630.318228] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113498, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.729569} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.318228] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7/dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 630.318228] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 630.318228] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-24400caa-ff8a-4375-b8c9-60e6863234b8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.326276] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Waiting for the task: (returnval){ [ 630.326276] env[65522]: value = "task-5113500" [ 630.326276] env[65522]: _type = "Task" [ 630.326276] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.338517] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113500, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.350406] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113499, 'name': ReconfigVM_Task, 'duration_secs': 0.336695} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.350406] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Reconfigured VM instance instance-00000001 to attach disk [datastore1] 57d32fb8-885e-44ba-967b-69f4bd1f5744/57d32fb8-885e-44ba-967b-69f4bd1f5744.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 630.351315] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-287fad4b-6dcf-47c1-8e6c-91916d03e586 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.358544] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 630.358544] env[65522]: value = "task-5113501" [ 630.358544] env[65522]: _type = "Task" [ 630.358544] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.377183] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5228f84d-455a-6cfc-c24a-9456dab0fcaf, 'name': SearchDatastore_Task, 'duration_secs': 0.059034} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.381185] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 630.381505] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] c93b52c3-1143-44d7-8c10-ac58bc25f541/c93b52c3-1143-44d7-8c10-ac58bc25f541.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 630.381892] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113501, 'name': Rename_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.382131] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.382645] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 630.382645] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89c7487c-dc67-4cd4-a0eb-c8f9845bc98f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.385670] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-825da940-805e-43f2-bba1-3e58eb163632 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.394833] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 630.394833] env[65522]: value = "task-5113502" [ 630.394833] env[65522]: _type = "Task" [ 630.394833] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.397602] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 630.397935] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 630.402284] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4f4b16a-b575-4d4a-adb1-14834856d18a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.409184] env[65522]: DEBUG nova.compute.utils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 630.422030] env[65522]: DEBUG nova.compute.manager [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 630.422030] env[65522]: DEBUG nova.network.neutron [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 630.422030] env[65522]: WARNING neutronclient.v2_0.client [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 630.422030] env[65522]: WARNING neutronclient.v2_0.client [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 630.422030] env[65522]: WARNING openstack [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 630.422420] env[65522]: WARNING openstack [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 630.429155] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113502, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.429486] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Waiting for the task: (returnval){ [ 630.429486] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529fc1a9-277b-efb0-1e84-cca55e3ea41a" [ 630.429486] env[65522]: _type = "Task" [ 630.429486] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.443808] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529fc1a9-277b-efb0-1e84-cca55e3ea41a, 'name': SearchDatastore_Task, 'duration_secs': 0.015375} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.445087] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08262625-2d7b-426c-97c2-ffff4f733fa4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.450957] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Waiting for the task: (returnval){ [ 630.450957] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b4f281-6f73-be19-3d5c-2e0eb29e6ffe" [ 630.450957] env[65522]: _type = "Task" [ 630.450957] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.459899] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b4f281-6f73-be19-3d5c-2e0eb29e6ffe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.538128] env[65522]: DEBUG nova.policy [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6dc2c46256514d9c87709cea348d64c5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '68c8ff7c2b604c939a2124142d1654b8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 630.578941] env[65522]: DEBUG nova.network.neutron [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Successfully updated port: 3896379b-7eb4-4dc5-819d-f68421a44dab {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 630.624704] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Releasing lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 630.626807] env[65522]: DEBUG nova.compute.manager [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Instance network_info: |[{"id": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "address": "fa:16:3e:9b:22:d2", "network": {"id": "b1e15367-b861-4a6c-8f94-921ed2b0a66f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1304672897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a47babe780ec4ae7b6c40f4531f09a38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "79c2e589-f55b-4843-8d99-2e565be16706", "external-id": "nsx-vlan-transportzone-858", "segmentation_id": 858, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcf081b2-7a", "ovs_interfaceid": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 630.628038] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9b:22:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '79c2e589-f55b-4843-8d99-2e565be16706', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dcf081b2-7a3e-4efd-84fa-bbf035dfd690', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 630.636138] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Creating folder: Project (a47babe780ec4ae7b6c40f4531f09a38). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 630.637380] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2bec8a35-1ba5-467e-b0bd-d140bb6b51d9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.650540] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Created folder: Project (a47babe780ec4ae7b6c40f4531f09a38) in parent group-v994660. [ 630.650770] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Creating folder: Instances. Parent ref: group-v994676. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 630.651072] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-effe0f7e-a116-4e88-945c-efd336569387 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.667225] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Created folder: Instances in parent group-v994676. [ 630.667495] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 630.670075] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 630.670075] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-50078b51-2c65-48ec-8651-a371018ed943 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.690553] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 630.690553] env[65522]: value = "task-5113505" [ 630.690553] env[65522]: _type = "Task" [ 630.690553] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.703685] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113505, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.839079] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113500, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085893} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.839493] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 630.840432] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-187fb01e-f949-4a05-9971-585ef9a1c8d6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.866149] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7/dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 630.866556] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9a25fe9-2366-48b5-aa9f-20d503396a79 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.891238] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113501, 'name': Rename_Task, 'duration_secs': 0.154265} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.892744] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 630.893115] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Waiting for the task: (returnval){ [ 630.893115] env[65522]: value = "task-5113506" [ 630.893115] env[65522]: _type = "Task" [ 630.893115] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.893711] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98b75563-8d1c-431e-b918-4332c7d0d05e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.909116] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 630.909116] env[65522]: value = "task-5113507" [ 630.909116] env[65522]: _type = "Task" [ 630.909116] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.913167] env[65522]: DEBUG nova.compute.manager [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 630.916690] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113506, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.925423] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113502, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.936625] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113507, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.965473] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b4f281-6f73-be19-3d5c-2e0eb29e6ffe, 'name': SearchDatastore_Task, 'duration_secs': 0.012501} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.966353] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 630.966772] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] fbb035b5-4562-44e3-9b95-8cf85299bb43/fbb035b5-4562-44e3-9b95-8cf85299bb43.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 630.969691] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.969960] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 630.970459] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36535ff7-19ec-48f1-8b84-c7af1a82d320 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.973299] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c5e9c5d-5750-4552-b78a-f9b1f61ff097 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.983051] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Waiting for the task: (returnval){ [ 630.983051] env[65522]: value = "task-5113508" [ 630.983051] env[65522]: _type = "Task" [ 630.983051] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.984122] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 630.984404] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 630.989056] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1018718c-a937-45c7-aa38-b5bc19db73c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.005643] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113508, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.006098] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 631.006098] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52eee552-76f9-2a7f-3114-d4f422206a15" [ 631.006098] env[65522]: _type = "Task" [ 631.006098] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.018649] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52eee552-76f9-2a7f-3114-d4f422206a15, 'name': SearchDatastore_Task, 'duration_secs': 0.012216} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.023478] env[65522]: DEBUG nova.network.neutron [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Successfully created port: 9902660c-976f-4cee-8338-74f077c8f7b3 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 631.025952] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-431f4f27-9afd-4786-8c35-d3fd297917ae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.032713] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 631.032713] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5280c761-b22d-5d55-03b2-b2d05bebd08e" [ 631.032713] env[65522]: _type = "Task" [ 631.032713] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.042440] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5280c761-b22d-5d55-03b2-b2d05bebd08e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.084412] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Acquiring lock "refresh_cache-1d3a9ada-df46-46af-a7f1-a84c251af4ac" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.085268] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Acquired lock "refresh_cache-1d3a9ada-df46-46af-a7f1-a84c251af4ac" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 631.085268] env[65522]: DEBUG nova.network.neutron [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 631.205885] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113505, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.211960] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fef134-1af0-4f1f-9d4a-611eadc64d45 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.222052] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd34908-4282-4223-8335-be789dae363b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.242224] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Received event network-changed-7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 631.242519] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Refreshing instance network info cache due to event network-changed-7d49f270-6cbd-4767-91fc-f899c46564c9. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 631.242790] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Acquiring lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.242944] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Acquired lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 631.243151] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Refreshing network info cache for port 7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 631.277856] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf5c406e-9fc2-4b19-bf5a-6a79bab68059 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.292659] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e722f86a-661c-425c-b9a4-654ab67c7597 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.320194] env[65522]: DEBUG nova.compute.provider_tree [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.410051] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113506, 'name': ReconfigVM_Task, 'duration_secs': 0.411146} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.411177] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Reconfigured VM instance instance-00000002 to attach disk [datastore1] dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7/dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 631.411810] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6b32ba2-2968-4272-81b2-5ff5a7338d69 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.418486] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113502, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587573} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.423099] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] c93b52c3-1143-44d7-8c10-ac58bc25f541/c93b52c3-1143-44d7-8c10-ac58bc25f541.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 631.423384] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 631.423686] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d59eca3e-88b8-42ea-a308-a98f3e817aaa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.437260] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Waiting for the task: (returnval){ [ 631.437260] env[65522]: value = "task-5113509" [ 631.437260] env[65522]: _type = "Task" [ 631.437260] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.437260] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113507, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.445894] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 631.445894] env[65522]: value = "task-5113510" [ 631.445894] env[65522]: _type = "Task" [ 631.445894] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.451200] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113509, 'name': Rename_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.462277] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113510, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.496140] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113508, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.546292] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5280c761-b22d-5d55-03b2-b2d05bebd08e, 'name': SearchDatastore_Task, 'duration_secs': 0.010655} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.550028] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 631.550028] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] b2d70982-54e7-459c-a0d8-48bf4b6e4345/b2d70982-54e7-459c-a0d8-48bf4b6e4345.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 631.550028] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-25084d61-0b87-4c7e-ad31-0f1eda5aa68a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.556551] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 631.556551] env[65522]: value = "task-5113511" [ 631.556551] env[65522]: _type = "Task" [ 631.556551] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.568286] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113511, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.588826] env[65522]: WARNING openstack [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.589143] env[65522]: WARNING openstack [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.686966] env[65522]: DEBUG nova.network.neutron [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 631.704602] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113505, 'name': CreateVM_Task, 'duration_secs': 0.657605} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.704772] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 631.705449] env[65522]: WARNING neutronclient.v2_0.client [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 631.706511] env[65522]: DEBUG oslo_vmware.service [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17bc33f2-2c7c-452b-9fb1-a7f0be4577d4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.712864] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.713023] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 631.713390] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 631.713640] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba4d1962-635f-4c37-8282-4518186a9c0c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.718142] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 631.718142] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5226ea8a-5341-4553-50e7-4d065b6f621f" [ 631.718142] env[65522]: _type = "Task" [ 631.718142] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.726353] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5226ea8a-5341-4553-50e7-4d065b6f621f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.776654] env[65522]: WARNING openstack [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.777171] env[65522]: WARNING openstack [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.785133] env[65522]: WARNING neutronclient.v2_0.client [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 631.785874] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.786385] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.819236] env[65522]: DEBUG nova.compute.manager [req-4ddb849a-23f2-49c5-9da6-5aad975968a5 req-38a09bbf-f428-4d6c-b94b-2797a0134258 service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Received event network-vif-plugged-dcf081b2-7a3e-4efd-84fa-bbf035dfd690 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 631.819449] env[65522]: DEBUG oslo_concurrency.lockutils [req-4ddb849a-23f2-49c5-9da6-5aad975968a5 req-38a09bbf-f428-4d6c-b94b-2797a0134258 service nova] Acquiring lock "6abf58ca-ad6b-4fe8-844a-ca33e62a844d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 631.819654] env[65522]: DEBUG oslo_concurrency.lockutils [req-4ddb849a-23f2-49c5-9da6-5aad975968a5 req-38a09bbf-f428-4d6c-b94b-2797a0134258 service nova] Lock "6abf58ca-ad6b-4fe8-844a-ca33e62a844d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 631.819817] env[65522]: DEBUG oslo_concurrency.lockutils [req-4ddb849a-23f2-49c5-9da6-5aad975968a5 req-38a09bbf-f428-4d6c-b94b-2797a0134258 service nova] Lock "6abf58ca-ad6b-4fe8-844a-ca33e62a844d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 631.819990] env[65522]: DEBUG nova.compute.manager [req-4ddb849a-23f2-49c5-9da6-5aad975968a5 req-38a09bbf-f428-4d6c-b94b-2797a0134258 service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] No waiting events found dispatching network-vif-plugged-dcf081b2-7a3e-4efd-84fa-bbf035dfd690 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 631.820152] env[65522]: WARNING nova.compute.manager [req-4ddb849a-23f2-49c5-9da6-5aad975968a5 req-38a09bbf-f428-4d6c-b94b-2797a0134258 service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Received unexpected event network-vif-plugged-dcf081b2-7a3e-4efd-84fa-bbf035dfd690 for instance with vm_state building and task_state spawning. [ 631.822610] env[65522]: DEBUG nova.scheduler.client.report [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 631.924168] env[65522]: WARNING neutronclient.v2_0.client [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 631.924849] env[65522]: WARNING openstack [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.925340] env[65522]: WARNING openstack [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.942405] env[65522]: DEBUG nova.compute.manager [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 631.945067] env[65522]: DEBUG oslo_vmware.api [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113507, 'name': PowerOnVM_Task, 'duration_secs': 0.602338} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.948847] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 631.949394] env[65522]: INFO nova.compute.manager [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Took 15.68 seconds to spawn the instance on the hypervisor. [ 631.949685] env[65522]: DEBUG nova.compute.manager [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 631.950801] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776ebc0a-f0bc-4ff5-929d-78ad3bd716e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.967353] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113509, 'name': Rename_Task, 'duration_secs': 0.267718} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.971789] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 631.975019] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0bb4ee54-70c4-446f-a0d2-a12b83b82c5e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.980512] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113510, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.23001} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.983149] env[65522]: DEBUG nova.virt.hardware [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 631.983411] env[65522]: DEBUG nova.virt.hardware [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 631.983572] env[65522]: DEBUG nova.virt.hardware [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 631.983773] env[65522]: DEBUG nova.virt.hardware [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 631.983933] env[65522]: DEBUG nova.virt.hardware [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 631.984088] env[65522]: DEBUG nova.virt.hardware [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 631.984297] env[65522]: DEBUG nova.virt.hardware [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 631.984451] env[65522]: DEBUG nova.virt.hardware [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 631.984638] env[65522]: DEBUG nova.virt.hardware [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 631.984788] env[65522]: DEBUG nova.virt.hardware [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 631.984977] env[65522]: DEBUG nova.virt.hardware [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 631.985806] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 631.986840] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f372bb72-8032-44f0-afa3-72551b4cb939 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.992191] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea0d17c9-eee3-4f1b-a902-1a4dfcd44225 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.999402] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Waiting for the task: (returnval){ [ 631.999402] env[65522]: value = "task-5113512" [ 631.999402] env[65522]: _type = "Task" [ 631.999402] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.019850] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113508, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.625978} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.030869] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Reconfiguring VM instance instance-00000003 to attach disk [datastore1] c93b52c3-1143-44d7-8c10-ac58bc25f541/c93b52c3-1143-44d7-8c10-ac58bc25f541.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 632.035380] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] fbb035b5-4562-44e3-9b95-8cf85299bb43/fbb035b5-4562-44e3-9b95-8cf85299bb43.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 632.035715] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 632.036046] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40fc8a9c-b054-4c97-88bc-117f9c4d13ae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.057627] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c9e2b747-4378-4e92-8af6-cc1fe0ea4040 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.060722] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0ca7202-e3e3-478d-8f44-f25f04e4d7c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.083090] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 632.083090] env[65522]: value = "task-5113513" [ 632.083090] env[65522]: _type = "Task" [ 632.083090] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.083376] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113512, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.097736] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Waiting for the task: (returnval){ [ 632.097736] env[65522]: value = "task-5113514" [ 632.097736] env[65522]: _type = "Task" [ 632.097736] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.113356] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113511, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497429} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.115169] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] b2d70982-54e7-459c-a0d8-48bf4b6e4345/b2d70982-54e7-459c-a0d8-48bf4b6e4345.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 632.115169] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 632.115709] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-56c6b38c-077c-45ec-bd9b-d18afd47360c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.128166] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113514, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.128166] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113513, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.137963] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 632.137963] env[65522]: value = "task-5113515" [ 632.137963] env[65522]: _type = "Task" [ 632.137963] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.152039] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113515, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.229294] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 632.229546] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 632.229790] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.229934] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 632.230127] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 632.230400] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e33e7510-d131-45b6-abd3-786f7c0831c3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.242820] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 632.243060] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 632.243917] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84557968-8797-460e-89e0-b43ea880d354 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.252431] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd955c9e-ada8-45ca-8e32-4c0d89243504 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.259239] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 632.259239] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526390fd-4108-8ff0-8d5b-af4032f5ae4c" [ 632.259239] env[65522]: _type = "Task" [ 632.259239] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.268968] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526390fd-4108-8ff0-8d5b-af4032f5ae4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.334500] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 632.335319] env[65522]: DEBUG nova.compute.manager [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 632.338421] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.590s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.340294] env[65522]: INFO nova.compute.claims [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.492595] env[65522]: INFO nova.compute.manager [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Took 22.10 seconds to build instance. [ 632.512009] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113512, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.556607] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.556607] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.612819] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113513, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.619711] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113514, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.123598} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.619891] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 632.620916] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9781a1c-be98-4407-bbc8-2a6083a77ded {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.656271] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Reconfiguring VM instance instance-00000004 to attach disk [datastore1] fbb035b5-4562-44e3-9b95-8cf85299bb43/fbb035b5-4562-44e3-9b95-8cf85299bb43.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 632.657676] env[65522]: DEBUG nova.network.neutron [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Updating instance_info_cache with network_info: [{"id": "3896379b-7eb4-4dc5-819d-f68421a44dab", "address": "fa:16:3e:c8:84:60", "network": {"id": "2dbba2da-d7e7-45b3-953b-aeb93ddb6177", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-737085066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "345c026cba1d40c59cab5aa5b477b833", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3896379b-7e", "ovs_interfaceid": "3896379b-7eb4-4dc5-819d-f68421a44dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 632.663736] env[65522]: DEBUG nova.network.neutron [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Successfully updated port: 9902660c-976f-4cee-8338-74f077c8f7b3 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 632.668502] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b72d1803-eb71-45bf-af98-919c513d439f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.686657] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Releasing lock "refresh_cache-1d3a9ada-df46-46af-a7f1-a84c251af4ac" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 632.687070] env[65522]: DEBUG nova.compute.manager [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Instance network_info: |[{"id": "3896379b-7eb4-4dc5-819d-f68421a44dab", "address": "fa:16:3e:c8:84:60", "network": {"id": "2dbba2da-d7e7-45b3-953b-aeb93ddb6177", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-737085066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "345c026cba1d40c59cab5aa5b477b833", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3896379b-7e", "ovs_interfaceid": "3896379b-7eb4-4dc5-819d-f68421a44dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 632.687810] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquiring lock "refresh_cache-bc40780e-43e2-4058-ab6f-d0d65e839128" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.688481] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquired lock "refresh_cache-bc40780e-43e2-4058-ab6f-d0d65e839128" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 632.688481] env[65522]: DEBUG nova.network.neutron [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 632.694481] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:84:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089a7624-43ba-4fce-bfc0-63e4bb7f9aeb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3896379b-7eb4-4dc5-819d-f68421a44dab', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 632.702422] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Creating folder: Project (345c026cba1d40c59cab5aa5b477b833). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 632.704648] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5651ce8b-56f8-4324-a75d-533b20ad41d2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.711007] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113515, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.175054} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.712948] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 632.713384] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Waiting for the task: (returnval){ [ 632.713384] env[65522]: value = "task-5113516" [ 632.713384] env[65522]: _type = "Task" [ 632.713384] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.717374] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9e5927-5791-49b2-b138-8b7e7a30b7e4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.725607] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Created folder: Project (345c026cba1d40c59cab5aa5b477b833) in parent group-v994660. [ 632.725872] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Creating folder: Instances. Parent ref: group-v994679. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 632.739069] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-55d67ff9-2671-45ec-9ce8-cb07b73b616e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.750535] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Reconfiguring VM instance instance-00000005 to attach disk [datastore1] b2d70982-54e7-459c-a0d8-48bf4b6e4345/b2d70982-54e7-459c-a0d8-48bf4b6e4345.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 632.754586] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-836f8ef4-8cdb-4351-b0f6-c0e0cc9f0622 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.770471] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113516, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.771972] env[65522]: WARNING neutronclient.v2_0.client [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 632.772452] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.772797] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.794167] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Preparing fetch location {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 632.794679] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Creating directory with path [datastore2] vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df/a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 632.794831] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Created folder: Instances in parent group-v994679. [ 632.795053] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 632.795311] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 632.795311] env[65522]: value = "task-5113519" [ 632.795311] env[65522]: _type = "Task" [ 632.795311] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.795505] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22e9c7de-5d71-4b41-8ab0-06aa4b684b96 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.797708] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 632.798024] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6c1e980f-17ba-4161-b8a9-ba1a66edb5b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.825214] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113519, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.826871] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 632.826871] env[65522]: value = "task-5113520" [ 632.826871] env[65522]: _type = "Task" [ 632.826871] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.836176] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113520, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.849521] env[65522]: DEBUG nova.compute.utils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 632.855320] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Created directory with path [datastore2] vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df/a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 632.855320] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Fetch image to [datastore2] vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 632.855320] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Downloading image file data a889db67-7337-4e32-8e34-642f34402926 to [datastore2] vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk on the data store datastore2 {{(pid=65522) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 632.855320] env[65522]: DEBUG nova.compute.manager [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 632.855649] env[65522]: DEBUG nova.network.neutron [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 632.855649] env[65522]: WARNING neutronclient.v2_0.client [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 632.855649] env[65522]: WARNING neutronclient.v2_0.client [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 632.856174] env[65522]: WARNING openstack [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.856494] env[65522]: WARNING openstack [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.863827] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3adf5861-992e-47ad-9754-f3cc101266ce {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.875677] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86176b4b-afc7-4132-a108-45b9f4dd711d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.888741] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0bae13-36d4-424d-98ab-ec8032a9db77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.939263] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Updated VIF entry in instance network info cache for port 7d49f270-6cbd-4767-91fc-f899c46564c9. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 632.939665] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Updating instance_info_cache with network_info: [{"id": "7d49f270-6cbd-4767-91fc-f899c46564c9", "address": "fa:16:3e:af:04:9f", "network": {"id": "6a60138b-7f03-4817-ac9a-1facdc1cc0cc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1344754077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4fdbd0ed78c41a591ccc0a587313ec8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f85835c8-5d0c-4b2f-97c4-6c4006580f79", "external-id": "nsx-vlan-transportzone-245", "segmentation_id": 245, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d49f270-6c", "ovs_interfaceid": "7d49f270-6cbd-4767-91fc-f899c46564c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 632.942272] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1e2092-930f-4ca1-8462-a4ec41e5408b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.953436] env[65522]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-30788fc8-8b6e-4b04-9caa-33080720c86d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.977621] env[65522]: DEBUG nova.policy [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a3e55b0cfeef4e3e9f1eef48f851a4a2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1927d35d9154557a33422045152daa6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 632.995891] env[65522]: DEBUG oslo_concurrency.lockutils [None req-79c0fd84-eb94-4557-b5fe-483f80afa9f5 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "57d32fb8-885e-44ba-967b-69f4bd1f5744" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.617s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 633.013708] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113512, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.057086] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Downloading image file data a889db67-7337-4e32-8e34-642f34402926 to the data store datastore2 {{(pid=65522) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 633.114883] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113513, 'name': ReconfigVM_Task, 'duration_secs': 0.664635} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.114883] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Reconfigured VM instance instance-00000003 to attach disk [datastore1] c93b52c3-1143-44d7-8c10-ac58bc25f541/c93b52c3-1143-44d7-8c10-ac58bc25f541.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 633.115213] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f4544f0-bbed-4b13-870b-1ffad2140d60 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.122246] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 633.122246] env[65522]: value = "task-5113521" [ 633.122246] env[65522]: _type = "Task" [ 633.122246] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.131586] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113521, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.132545] env[65522]: DEBUG oslo_vmware.rw_handles [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=65522) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 633.205087] env[65522]: WARNING openstack [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.205087] env[65522]: WARNING openstack [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.227933] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113516, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.309325] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113519, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.340205] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113520, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.355250] env[65522]: DEBUG nova.compute.manager [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 633.448800] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Releasing lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 633.449102] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Received event network-vif-plugged-e6e22697-3ece-4047-b378-b47c9c5de778 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 633.449311] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Acquiring lock "dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 633.449506] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Lock "dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 633.449660] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Lock "dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 633.450400] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] No waiting events found dispatching network-vif-plugged-e6e22697-3ece-4047-b378-b47c9c5de778 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 633.450400] env[65522]: WARNING nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Received unexpected event network-vif-plugged-e6e22697-3ece-4047-b378-b47c9c5de778 for instance with vm_state building and task_state spawning. [ 633.451059] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Received event network-changed-e6e22697-3ece-4047-b378-b47c9c5de778 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 633.451059] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Refreshing instance network info cache due to event network-changed-e6e22697-3ece-4047-b378-b47c9c5de778. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 633.451201] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Acquiring lock "refresh_cache-dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.451293] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Acquired lock "refresh_cache-dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.451453] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Refreshing network info cache for port e6e22697-3ece-4047-b378-b47c9c5de778 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 633.480315] env[65522]: DEBUG nova.network.neutron [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Successfully created port: a002c980-b974-432b-aa4e-e8613c617e8e {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 633.501906] env[65522]: DEBUG nova.compute.manager [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 633.520625] env[65522]: DEBUG oslo_vmware.api [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113512, 'name': PowerOnVM_Task, 'duration_secs': 1.025377} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.520953] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 633.521099] env[65522]: INFO nova.compute.manager [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Took 15.05 seconds to spawn the instance on the hypervisor. [ 633.521274] env[65522]: DEBUG nova.compute.manager [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 633.522144] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e39020-bcd4-424d-a30e-2364b371cd6a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.584726] env[65522]: DEBUG nova.network.neutron [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 633.641216] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113521, 'name': Rename_Task, 'duration_secs': 0.283447} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.642953] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 633.642953] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9e64a35c-5d30-4282-bad9-5889fbbd273d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.652558] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 633.652558] env[65522]: value = "task-5113522" [ 633.652558] env[65522]: _type = "Task" [ 633.652558] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.673106] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113522, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.702049] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03eaf33-9ca0-4cb9-97ae-94f1c94071b7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.710549] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43cbc8b5-9c8b-478c-9a34-58eef0fd5dd9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.749033] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c0b842-6062-4e65-9c6d-c58d9e918fe7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.762931] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113516, 'name': ReconfigVM_Task, 'duration_secs': 0.862369} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.764618] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da423283-2780-4fb1-9184-b61f8e3ebee5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.772167] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Reconfigured VM instance instance-00000004 to attach disk [datastore1] fbb035b5-4562-44e3-9b95-8cf85299bb43/fbb035b5-4562-44e3-9b95-8cf85299bb43.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 633.772167] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a478c94-f204-4302-95f3-d3444b886d25 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.784705] env[65522]: DEBUG nova.compute.provider_tree [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.790370] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Waiting for the task: (returnval){ [ 633.790370] env[65522]: value = "task-5113523" [ 633.790370] env[65522]: _type = "Task" [ 633.790370] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.812386] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113519, 'name': ReconfigVM_Task, 'duration_secs': 0.711154} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.812681] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Reconfigured VM instance instance-00000005 to attach disk [datastore1] b2d70982-54e7-459c-a0d8-48bf4b6e4345/b2d70982-54e7-459c-a0d8-48bf4b6e4345.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 633.814041] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ceccdbc-8fad-4750-86b5-8cfbb8a371fe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.823473] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 633.823473] env[65522]: value = "task-5113524" [ 633.823473] env[65522]: _type = "Task" [ 633.823473] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.835662] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113524, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.843156] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113520, 'name': CreateVM_Task, 'duration_secs': 0.697741} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.843156] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 633.843361] env[65522]: WARNING neutronclient.v2_0.client [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.843818] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.844013] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.844382] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 633.844759] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db5d6028-68e2-4b5d-a052-d586e6a6bc52 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.854200] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 633.854200] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5295b427-2b07-e57c-6a21-ffcd5b8e4eb1" [ 633.854200] env[65522]: _type = "Task" [ 633.854200] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.868574] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5295b427-2b07-e57c-6a21-ffcd5b8e4eb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.957857] env[65522]: WARNING neutronclient.v2_0.client [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.958512] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.958911] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.992511] env[65522]: WARNING openstack [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.993010] env[65522]: WARNING openstack [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 634.034738] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 634.048467] env[65522]: INFO nova.compute.manager [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Took 23.19 seconds to build instance. [ 634.164924] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113522, 'name': PowerOnVM_Task} progress is 92%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.205787] env[65522]: DEBUG oslo_vmware.rw_handles [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Completed reading data from the image iterator. {{(pid=65522) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 634.205787] env[65522]: DEBUG oslo_vmware.rw_handles [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 634.267602] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Downloaded image file data a889db67-7337-4e32-8e34-642f34402926 to vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk on the data store datastore2 {{(pid=65522) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 634.269494] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Caching image {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 634.269813] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Copying Virtual Disk [datastore2] vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk to [datastore2] vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 634.270164] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-289a8c5d-2501-473a-8ed4-2e08b2bb17ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.283158] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 634.283158] env[65522]: value = "task-5113525" [ 634.283158] env[65522]: _type = "Task" [ 634.283158] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.291808] env[65522]: DEBUG nova.scheduler.client.report [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 634.302373] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113525, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.308932] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113523, 'name': Rename_Task, 'duration_secs': 0.216149} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.309247] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 634.309492] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-45175852-787e-490b-b740-cf7ffd4435a1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.318287] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Waiting for the task: (returnval){ [ 634.318287] env[65522]: value = "task-5113526" [ 634.318287] env[65522]: _type = "Task" [ 634.318287] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.332886] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113526, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.336451] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113524, 'name': Rename_Task, 'duration_secs': 0.189378} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.336722] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 634.337025] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3563f2c1-1c91-4d39-938e-df8020569070 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.344224] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 634.344224] env[65522]: value = "task-5113527" [ 634.344224] env[65522]: _type = "Task" [ 634.344224] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.353205] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113527, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.366812] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.367171] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 634.369085] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.371271] env[65522]: DEBUG nova.compute.manager [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 634.406384] env[65522]: DEBUG nova.virt.hardware [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 634.406730] env[65522]: DEBUG nova.virt.hardware [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 634.406890] env[65522]: DEBUG nova.virt.hardware [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 634.407092] env[65522]: DEBUG nova.virt.hardware [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 634.407294] env[65522]: DEBUG nova.virt.hardware [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 634.407472] env[65522]: DEBUG nova.virt.hardware [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 634.407760] env[65522]: DEBUG nova.virt.hardware [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 634.407962] env[65522]: DEBUG nova.virt.hardware [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 634.408158] env[65522]: DEBUG nova.virt.hardware [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 634.408348] env[65522]: DEBUG nova.virt.hardware [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 634.408554] env[65522]: DEBUG nova.virt.hardware [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 634.409572] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee84d30-c9fe-4eab-aa10-c9a5ce1fc07d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.419198] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6b028d-32a2-4af0-a952-26f3756a6319 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.550867] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c281f443-405d-48a0-a7f7-668cdabee9bb tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Lock "dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.705s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 634.669270] env[65522]: DEBUG oslo_vmware.api [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113522, 'name': PowerOnVM_Task, 'duration_secs': 0.640361} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.669270] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 634.669270] env[65522]: INFO nova.compute.manager [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Took 13.86 seconds to spawn the instance on the hypervisor. [ 634.669270] env[65522]: DEBUG nova.compute.manager [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 634.669270] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e67842-6c2f-4301-a1e9-f8cdae316bb1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.794392] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113525, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.804972] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 634.805536] env[65522]: DEBUG nova.compute.manager [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 634.809435] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.775s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 634.810976] env[65522]: INFO nova.compute.claims [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 634.829092] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113526, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.859018] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113527, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.883915] env[65522]: WARNING neutronclient.v2_0.client [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 634.884688] env[65522]: WARNING openstack [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 634.884994] env[65522]: WARNING openstack [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.054895] env[65522]: DEBUG nova.compute.manager [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 635.192559] env[65522]: INFO nova.compute.manager [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Took 23.43 seconds to build instance. [ 635.192559] env[65522]: DEBUG nova.network.neutron [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Successfully updated port: a002c980-b974-432b-aa4e-e8613c617e8e {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 635.207278] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.207680] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.295597] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113525, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.965938} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.301311] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Copied Virtual Disk [datastore2] vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk to [datastore2] vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 635.301488] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Deleting the datastore file [datastore2] vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df/a889db67-7337-4e32-8e34-642f34402926/tmp-sparse.vmdk {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 635.301895] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d0a3ce6-23d1-4297-bce8-816614fdd1d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.311039] env[65522]: DEBUG nova.compute.utils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 635.313336] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 635.313336] env[65522]: value = "task-5113528" [ 635.313336] env[65522]: _type = "Task" [ 635.313336] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.313499] env[65522]: DEBUG nova.compute.manager [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 635.313597] env[65522]: DEBUG nova.network.neutron [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 635.314643] env[65522]: WARNING neutronclient.v2_0.client [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.314643] env[65522]: WARNING neutronclient.v2_0.client [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.317699] env[65522]: WARNING openstack [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.317699] env[65522]: WARNING openstack [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.345566] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113528, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.345889] env[65522]: DEBUG oslo_vmware.api [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113526, 'name': PowerOnVM_Task, 'duration_secs': 0.691182} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.347066] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 635.347286] env[65522]: INFO nova.compute.manager [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Took 12.30 seconds to spawn the instance on the hypervisor. [ 635.347536] env[65522]: DEBUG nova.compute.manager [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 635.350815] env[65522]: DEBUG nova.network.neutron [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Updating instance_info_cache with network_info: [{"id": "9902660c-976f-4cee-8338-74f077c8f7b3", "address": "fa:16:3e:20:4e:b5", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.243", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9902660c-97", "ovs_interfaceid": "9902660c-976f-4cee-8338-74f077c8f7b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 635.361832] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e941e1b0-016b-453e-8dea-0e9b1bae8acd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.367740] env[65522]: WARNING neutronclient.v2_0.client [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.368965] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.369392] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.386330] env[65522]: DEBUG oslo_vmware.api [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113527, 'name': PowerOnVM_Task, 'duration_secs': 0.642178} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.388440] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 635.388440] env[65522]: INFO nova.compute.manager [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Took 10.54 seconds to spawn the instance on the hypervisor. [ 635.388440] env[65522]: DEBUG nova.compute.manager [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 635.393026] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e675d155-1f03-443d-aa46-f4dc5e21e275 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.412916] env[65522]: DEBUG nova.policy [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18ceaff46b314c9f9da68a8d0c6f5f2b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fa11b46d9fe144f391233e6eb9c819d7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 635.513336] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Updated VIF entry in instance network info cache for port e6e22697-3ece-4047-b378-b47c9c5de778. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 635.513700] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Updating instance_info_cache with network_info: [{"id": "e6e22697-3ece-4047-b378-b47c9c5de778", "address": "fa:16:3e:1a:4e:a7", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.26", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6e22697-3e", "ovs_interfaceid": "e6e22697-3ece-4047-b378-b47c9c5de778", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 635.583890] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 635.694735] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50be527c-537f-4eed-95f1-e210c260ee2a tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "c93b52c3-1143-44d7-8c10-ac58bc25f541" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.948s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.695357] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Acquiring lock "refresh_cache-a09c12be-0aaa-4934-a1b3-5af29feba8a6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.695468] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Acquired lock "refresh_cache-a09c12be-0aaa-4934-a1b3-5af29feba8a6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.695708] env[65522]: DEBUG nova.network.neutron [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 635.815300] env[65522]: DEBUG nova.compute.manager [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 635.836914] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113528, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.041089} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.837187] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 635.837512] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Moving file from [datastore2] vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df/a889db67-7337-4e32-8e34-642f34402926 to [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926. {{(pid=65522) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 635.837866] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-fd6d7bc9-6eb7-4a57-a3f1-e5158649cf11 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.847405] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 635.847405] env[65522]: value = "task-5113529" [ 635.847405] env[65522]: _type = "Task" [ 635.847405] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.853859] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Releasing lock "refresh_cache-bc40780e-43e2-4058-ab6f-d0d65e839128" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 635.854297] env[65522]: DEBUG nova.compute.manager [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Instance network_info: |[{"id": "9902660c-976f-4cee-8338-74f077c8f7b3", "address": "fa:16:3e:20:4e:b5", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.243", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9902660c-97", "ovs_interfaceid": "9902660c-976f-4cee-8338-74f077c8f7b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 635.861199] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:4e:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9902660c-976f-4cee-8338-74f077c8f7b3', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 635.869351] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Creating folder: Project (68c8ff7c2b604c939a2124142d1654b8). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 635.869653] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113529, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.870910] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0042d08c-1937-42aa-8294-b874d2fbbecf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.884689] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Created folder: Project (68c8ff7c2b604c939a2124142d1654b8) in parent group-v994660. [ 635.884906] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Creating folder: Instances. Parent ref: group-v994682. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 635.888205] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3332e490-6bfb-4307-97fc-296ab13d93ff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.898280] env[65522]: DEBUG nova.network.neutron [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Successfully created port: cf16c93a-755c-435b-8c86-833d8ea8afbf {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 635.902472] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Created folder: Instances in parent group-v994682. [ 635.903224] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 635.907096] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 635.911679] env[65522]: INFO nova.compute.manager [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Took 21.87 seconds to build instance. [ 635.912966] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8932ad3c-cd07-429e-85f8-a21301d6cc85 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.930901] env[65522]: INFO nova.compute.manager [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Took 17.74 seconds to build instance. [ 635.943622] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 635.943622] env[65522]: value = "task-5113532" [ 635.943622] env[65522]: _type = "Task" [ 635.943622] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.954065] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113532, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.018969] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Releasing lock "refresh_cache-dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 636.019338] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Received event network-vif-plugged-573a1035-e00a-4cd2-a067-934487b65b60 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 636.019459] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Acquiring lock "c93b52c3-1143-44d7-8c10-ac58bc25f541-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 636.019626] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Lock "c93b52c3-1143-44d7-8c10-ac58bc25f541-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 636.019787] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Lock "c93b52c3-1143-44d7-8c10-ac58bc25f541-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 636.019990] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] No waiting events found dispatching network-vif-plugged-573a1035-e00a-4cd2-a067-934487b65b60 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 636.020180] env[65522]: WARNING nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Received unexpected event network-vif-plugged-573a1035-e00a-4cd2-a067-934487b65b60 for instance with vm_state building and task_state spawning. [ 636.020368] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Received event network-changed-573a1035-e00a-4cd2-a067-934487b65b60 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 636.020601] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Refreshing instance network info cache due to event network-changed-573a1035-e00a-4cd2-a067-934487b65b60. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 636.020878] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Acquiring lock "refresh_cache-c93b52c3-1143-44d7-8c10-ac58bc25f541" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.021406] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Acquired lock "refresh_cache-c93b52c3-1143-44d7-8c10-ac58bc25f541" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 636.021406] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Refreshing network info cache for port 573a1035-e00a-4cd2-a067-934487b65b60 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 636.100943] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9332b089-cd32-4865-a4ba-0faf9a72d2c2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.110315] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd67666c-138f-4567-bc69-16ae72b11d0b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.151893] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67193fb-2f1f-4258-91b1-784a96e3ad87 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.161298] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a7ae03c-3271-4d2e-ad89-fc2ca8c756c3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.177384] env[65522]: DEBUG nova.compute.provider_tree [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 636.203746] env[65522]: DEBUG nova.compute.manager [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 636.206875] env[65522]: WARNING openstack [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.208803] env[65522]: WARNING openstack [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.362729] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113529, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.039651} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.363036] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] File moved {{(pid=65522) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 636.363236] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Cleaning up location [datastore2] vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 636.363413] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Deleting the datastore file [datastore2] vmware_temp/4d747c1d-9c23-4961-b998-fe16927628df {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 636.363656] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6cc2e3ef-cf78-49f4-a459-ed2ea45688b8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.372330] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 636.372330] env[65522]: value = "task-5113533" [ 636.372330] env[65522]: _type = "Task" [ 636.372330] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.381981] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113533, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.430743] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40fb1f7e-c128-4e54-ab79-6cbe284af7a1 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Lock "fbb035b5-4562-44e3-9b95-8cf85299bb43" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.401s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 636.439083] env[65522]: DEBUG oslo_concurrency.lockutils [None req-36c12146-2260-4a24-a5a3-873e0b44fc45 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.260s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 636.454617] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113532, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.527026] env[65522]: WARNING neutronclient.v2_0.client [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 636.527026] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.527026] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.554540] env[65522]: DEBUG nova.network.neutron [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 636.701468] env[65522]: ERROR nova.scheduler.client.report [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [req-aa2d7b57-68f3-4cf0-81da-53ab9208c5c6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9a0d1ac8-ed26-4657-8899-fbceab29db0d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-aa2d7b57-68f3-4cf0-81da-53ab9208c5c6"}]} [ 636.722523] env[65522]: DEBUG nova.scheduler.client.report [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Refreshing inventories for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 636.738715] env[65522]: DEBUG nova.scheduler.client.report [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Updating ProviderTree inventory for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 636.739106] env[65522]: DEBUG nova.compute.provider_tree [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 636.742032] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 636.750822] env[65522]: DEBUG nova.scheduler.client.report [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Refreshing aggregate associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, aggregates: None {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 636.769452] env[65522]: DEBUG nova.scheduler.client.report [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Refreshing trait associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 636.828251] env[65522]: DEBUG nova.compute.manager [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 636.855313] env[65522]: DEBUG nova.virt.hardware [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 636.855573] env[65522]: DEBUG nova.virt.hardware [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 636.855760] env[65522]: DEBUG nova.virt.hardware [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 636.855961] env[65522]: DEBUG nova.virt.hardware [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 636.856112] env[65522]: DEBUG nova.virt.hardware [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 636.856255] env[65522]: DEBUG nova.virt.hardware [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 636.856458] env[65522]: DEBUG nova.virt.hardware [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.856612] env[65522]: DEBUG nova.virt.hardware [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 636.856790] env[65522]: DEBUG nova.virt.hardware [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 636.856919] env[65522]: DEBUG nova.virt.hardware [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 636.857111] env[65522]: DEBUG nova.virt.hardware [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 636.858044] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514e5fc1-ea3c-4bc2-bc05-4b318d64c9cc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.868705] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050b6d36-cc34-47f7-9863-30eea5988bab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.893837] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113533, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.062367} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.894232] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 636.894946] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c44079ec-62d3-4aba-9694-e9a314ebd557 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.900725] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 636.900725] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52898ff1-1b5c-8a3a-72c6-1892cbcd601f" [ 636.900725] env[65522]: _type = "Task" [ 636.900725] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.912916] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52898ff1-1b5c-8a3a-72c6-1892cbcd601f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.931414] env[65522]: WARNING openstack [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.931783] env[65522]: WARNING openstack [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.960033] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113532, 'name': CreateVM_Task, 'duration_secs': 0.598185} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.960226] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 636.960722] env[65522]: WARNING neutronclient.v2_0.client [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 636.961107] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.961248] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 636.961558] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 636.961827] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5035db06-ca06-4bd2-bf8f-1aa49631e28d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.967447] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 636.967447] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529e2edb-1156-1f29-d857-b30085250604" [ 636.967447] env[65522]: _type = "Task" [ 636.967447] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.969222] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-481d56b7-3aa0-4129-bdf3-546d1e416c53 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.982720] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529e2edb-1156-1f29-d857-b30085250604, 'name': SearchDatastore_Task, 'duration_secs': 0.011578} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.983438] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 636.984020] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 636.984020] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.984966] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b7762e5-dc6d-4013-81e4-1ee0d655d5eb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.021076] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72063e11-22cc-4d5a-b4e3-dc7220fd82e1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.029498] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6524f897-f29a-4198-a71c-7c1f4c9058ab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.043911] env[65522]: DEBUG nova.compute.provider_tree [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 637.412311] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52898ff1-1b5c-8a3a-72c6-1892cbcd601f, 'name': SearchDatastore_Task, 'duration_secs': 0.011718} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.412860] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 637.412860] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 6abf58ca-ad6b-4fe8-844a-ca33e62a844d/6abf58ca-ad6b-4fe8-844a-ca33e62a844d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 637.413193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 637.413426] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 637.413630] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f904e6a0-70ed-40bc-822b-70e9910a3d75 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.416014] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-adcebf20-ce01-47b2-9d43-a6bbfc2cbbfe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.423923] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 637.423923] env[65522]: value = "task-5113534" [ 637.423923] env[65522]: _type = "Task" [ 637.423923] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.428337] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 637.428520] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 637.430225] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec464165-29e4-4cae-91f0-cdb4c27ce769 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.436967] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113534, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.440954] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 637.440954] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f93738-eb41-2142-1869-8d2c67e565ac" [ 637.440954] env[65522]: _type = "Task" [ 637.440954] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.450953] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f93738-eb41-2142-1869-8d2c67e565ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.544717] env[65522]: DEBUG nova.network.neutron [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Successfully updated port: cf16c93a-755c-435b-8c86-833d8ea8afbf {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 637.588775] env[65522]: DEBUG nova.scheduler.client.report [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Updated inventory for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with generation 17 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 637.589116] env[65522]: DEBUG nova.compute.provider_tree [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Updating resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d generation from 17 to 18 during operation: update_inventory {{(pid=65522) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 637.589620] env[65522]: DEBUG nova.compute.provider_tree [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 637.902383] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.902773] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.932877] env[65522]: WARNING neutronclient.v2_0.client [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 637.933693] env[65522]: WARNING openstack [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.934136] env[65522]: WARNING openstack [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.949464] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113534, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.956914] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f93738-eb41-2142-1869-8d2c67e565ac, 'name': SearchDatastore_Task, 'duration_secs': 0.016897} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.957615] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-543124f9-13d0-4795-8c36-f41a3e420760 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.965790] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 637.965790] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529bee67-8c19-fa3b-d648-e0b8a1e28f8a" [ 637.965790] env[65522]: _type = "Task" [ 637.965790] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.977483] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529bee67-8c19-fa3b-d648-e0b8a1e28f8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.048754] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.049034] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.049269] env[65522]: DEBUG nova.network.neutron [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 638.095316] env[65522]: DEBUG nova.compute.manager [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Received event network-changed-dcf081b2-7a3e-4efd-84fa-bbf035dfd690 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 638.095498] env[65522]: DEBUG nova.compute.manager [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Refreshing instance network info cache due to event network-changed-dcf081b2-7a3e-4efd-84fa-bbf035dfd690. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 638.095714] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Acquiring lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.095844] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Acquired lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.095992] env[65522]: DEBUG nova.network.neutron [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Refreshing network info cache for port dcf081b2-7a3e-4efd-84fa-bbf035dfd690 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 638.101021] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.289s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 638.101021] env[65522]: DEBUG nova.compute.manager [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 638.101491] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.518s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 638.102932] env[65522]: INFO nova.compute.claims [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 638.435192] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113534, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.65044} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.435485] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 6abf58ca-ad6b-4fe8-844a-ca33e62a844d/6abf58ca-ad6b-4fe8-844a-ca33e62a844d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 638.436202] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 638.436202] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b691fe8-d5ec-489c-aed3-0bcb888eaf56 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.440401] env[65522]: WARNING neutronclient.v2_0.client [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.440666] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.441562] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.451786] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 638.451786] env[65522]: value = "task-5113535" [ 638.451786] env[65522]: _type = "Task" [ 638.451786] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.461697] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113535, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.476634] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529bee67-8c19-fa3b-d648-e0b8a1e28f8a, 'name': SearchDatastore_Task, 'duration_secs': 0.058692} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.476951] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.477263] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 1d3a9ada-df46-46af-a7f1-a84c251af4ac/1d3a9ada-df46-46af-a7f1-a84c251af4ac.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 638.477581] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.477881] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 638.478159] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-62ab4926-ea0d-4b80-af5d-25a50f8032ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.481775] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9ef65aad-b8c7-4901-8903-52eb50727c72 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.490683] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 638.490683] env[65522]: value = "task-5113536" [ 638.490683] env[65522]: _type = "Task" [ 638.490683] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.497529] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 638.497819] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 638.499052] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17bb8c66-8a45-4399-acbe-db41192102d9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.506801] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113536, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.511241] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 638.511241] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5242a71a-9638-4962-e227-0dfd87c71f95" [ 638.511241] env[65522]: _type = "Task" [ 638.511241] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.521199] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5242a71a-9638-4962-e227-0dfd87c71f95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.552435] env[65522]: WARNING openstack [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.552929] env[65522]: WARNING openstack [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.562534] env[65522]: DEBUG nova.network.neutron [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Updating instance_info_cache with network_info: [{"id": "a002c980-b974-432b-aa4e-e8613c617e8e", "address": "fa:16:3e:d4:83:9e", "network": {"id": "65a80e4b-308a-433d-8c2d-4f21879d8ac2", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1688898810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1927d35d9154557a33422045152daa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f847601f-7479-48eb-842f-41f94eea8537", "external-id": "nsx-vlan-transportzone-35", "segmentation_id": 35, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa002c980-b9", "ovs_interfaceid": "a002c980-b974-432b-aa4e-e8613c617e8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 638.602285] env[65522]: WARNING neutronclient.v2_0.client [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.605541] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.605541] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.614323] env[65522]: DEBUG nova.compute.utils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 638.621019] env[65522]: DEBUG nova.compute.manager [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 638.621019] env[65522]: DEBUG nova.network.neutron [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 638.621019] env[65522]: WARNING neutronclient.v2_0.client [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.621019] env[65522]: WARNING neutronclient.v2_0.client [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.621019] env[65522]: WARNING openstack [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.624859] env[65522]: WARNING openstack [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.811790] env[65522]: DEBUG nova.network.neutron [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 638.856927] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41b4d94-0c7a-4491-8ff9-b4069f7bb925 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.865390] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e319c14-50c6-4958-96d1-d645f62d7f5e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.900716] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Updated VIF entry in instance network info cache for port 573a1035-e00a-4cd2-a067-934487b65b60. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 638.901077] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Updating instance_info_cache with network_info: [{"id": "573a1035-e00a-4cd2-a067-934487b65b60", "address": "fa:16:3e:8a:04:eb", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap573a1035-e0", "ovs_interfaceid": "573a1035-e00a-4cd2-a067-934487b65b60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 638.905242] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7599c61-b527-46d7-9d1d-de4a03a97bb7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.912429] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c170d6-ca06-46e2-af47-e24426a02327 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.927736] env[65522]: DEBUG nova.compute.provider_tree [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.931681] env[65522]: WARNING openstack [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.931681] env[65522]: WARNING openstack [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.963101] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113535, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079135} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.963389] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 638.964290] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a83250-f801-4865-af45-bb9de00379db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.995171] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Reconfiguring VM instance instance-00000007 to attach disk [datastore2] 6abf58ca-ad6b-4fe8-844a-ca33e62a844d/6abf58ca-ad6b-4fe8-844a-ca33e62a844d.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 638.999743] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8419ece6-59d8-4314-ada3-57a737e61514 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.036913] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113536, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.037764] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5242a71a-9638-4962-e227-0dfd87c71f95, 'name': SearchDatastore_Task, 'duration_secs': 0.051846} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.040610] env[65522]: DEBUG nova.policy [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c42fe34a9211470f9ba22ab609cd782e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e099bf52db6241d78a15fed1c5bbfbc8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 639.045026] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 639.045026] env[65522]: value = "task-5113537" [ 639.045026] env[65522]: _type = "Task" [ 639.045026] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.045955] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d67c8a4a-84be-43e1-8db3-e957070858f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.059095] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 639.059095] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dcf617-d016-6a28-d1e4-10590350c0cd" [ 639.059095] env[65522]: _type = "Task" [ 639.059095] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.059386] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113537, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.065397] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Releasing lock "refresh_cache-a09c12be-0aaa-4934-a1b3-5af29feba8a6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.065798] env[65522]: DEBUG nova.compute.manager [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Instance network_info: |[{"id": "a002c980-b974-432b-aa4e-e8613c617e8e", "address": "fa:16:3e:d4:83:9e", "network": {"id": "65a80e4b-308a-433d-8c2d-4f21879d8ac2", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1688898810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1927d35d9154557a33422045152daa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f847601f-7479-48eb-842f-41f94eea8537", "external-id": "nsx-vlan-transportzone-35", "segmentation_id": 35, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa002c980-b9", "ovs_interfaceid": "a002c980-b974-432b-aa4e-e8613c617e8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 639.069621] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:83:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f847601f-7479-48eb-842f-41f94eea8537', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a002c980-b974-432b-aa4e-e8613c617e8e', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 639.077211] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Creating folder: Project (b1927d35d9154557a33422045152daa6). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 639.079021] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.079350] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.086763] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fba0d617-d3fd-40e9-b410-a20f56f95290 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.100453] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dcf617-d016-6a28-d1e4-10590350c0cd, 'name': SearchDatastore_Task, 'duration_secs': 0.012244} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.102189] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.102506] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] bc40780e-43e2-4058-ab6f-d0d65e839128/bc40780e-43e2-4058-ab6f-d0d65e839128.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 639.102762] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Created folder: Project (b1927d35d9154557a33422045152daa6) in parent group-v994660. [ 639.102925] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Creating folder: Instances. Parent ref: group-v994685. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 639.103173] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aaf543fa-98cd-419b-8eb6-8afd9b821f8d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.107103] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f603c8d7-affc-483a-a329-88a89e0c9dc4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.114044] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 639.114044] env[65522]: value = "task-5113540" [ 639.114044] env[65522]: _type = "Task" [ 639.114044] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.119577] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Created folder: Instances in parent group-v994685. [ 639.119852] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 639.120610] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 639.120837] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ffa0503e-8680-4363-a88e-764d9700fe50 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.141023] env[65522]: DEBUG nova.compute.manager [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 639.142559] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113540, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.149380] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 639.149380] env[65522]: value = "task-5113541" [ 639.149380] env[65522]: _type = "Task" [ 639.149380] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.158693] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113541, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.185321] env[65522]: WARNING neutronclient.v2_0.client [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 639.185321] env[65522]: WARNING openstack [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.185494] env[65522]: WARNING openstack [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.407434] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Releasing lock "refresh_cache-c93b52c3-1143-44d7-8c10-ac58bc25f541" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.407826] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Received event network-vif-plugged-43c29c73-562e-48ba-886e-8e42d9c04074 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 639.408504] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Acquiring lock "fbb035b5-4562-44e3-9b95-8cf85299bb43-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 639.408504] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Lock "fbb035b5-4562-44e3-9b95-8cf85299bb43-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 639.408504] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Lock "fbb035b5-4562-44e3-9b95-8cf85299bb43-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 639.408716] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] No waiting events found dispatching network-vif-plugged-43c29c73-562e-48ba-886e-8e42d9c04074 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 639.408833] env[65522]: WARNING nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Received unexpected event network-vif-plugged-43c29c73-562e-48ba-886e-8e42d9c04074 for instance with vm_state building and task_state spawning. [ 639.409070] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Received event network-changed-43c29c73-562e-48ba-886e-8e42d9c04074 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 639.409243] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Refreshing instance network info cache due to event network-changed-43c29c73-562e-48ba-886e-8e42d9c04074. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 639.409438] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Acquiring lock "refresh_cache-fbb035b5-4562-44e3-9b95-8cf85299bb43" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.410501] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Acquired lock "refresh_cache-fbb035b5-4562-44e3-9b95-8cf85299bb43" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.410501] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Refreshing network info cache for port 43c29c73-562e-48ba-886e-8e42d9c04074 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 639.440180] env[65522]: DEBUG nova.scheduler.client.report [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 639.512592] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113536, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.979004} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.513018] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 1d3a9ada-df46-46af-a7f1-a84c251af4ac/1d3a9ada-df46-46af-a7f1-a84c251af4ac.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 639.513362] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 639.513712] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b4cd3989-83e0-4e18-97c4-072f8e3d990f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.522920] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 639.522920] env[65522]: value = "task-5113542" [ 639.522920] env[65522]: _type = "Task" [ 639.522920] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.530181] env[65522]: DEBUG nova.network.neutron [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Successfully created port: 366aac9d-7922-4057-b141-5f4d54f4e2ab {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 639.536666] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113542, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.561610] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113537, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.631924] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113540, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.664716] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113541, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.916938] env[65522]: WARNING neutronclient.v2_0.client [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 639.918242] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.918708] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.946836] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.845s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 639.947508] env[65522]: DEBUG nova.compute.manager [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 639.951055] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.209s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 639.953715] env[65522]: INFO nova.compute.claims [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.034646] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113542, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093378} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.034920] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 640.035848] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea4710c-3b33-4488-a5ea-60ec92d5ac8f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.061107] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Reconfiguring VM instance instance-00000006 to attach disk [datastore2] 1d3a9ada-df46-46af-a7f1-a84c251af4ac/1d3a9ada-df46-46af-a7f1-a84c251af4ac.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 640.064657] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99d9fd38-0928-402a-bea8-3f0acdc492cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.087479] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113537, 'name': ReconfigVM_Task, 'duration_secs': 0.970127} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.088623] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Reconfigured VM instance instance-00000007 to attach disk [datastore2] 6abf58ca-ad6b-4fe8-844a-ca33e62a844d/6abf58ca-ad6b-4fe8-844a-ca33e62a844d.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 640.089437] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 640.089437] env[65522]: value = "task-5113543" [ 640.089437] env[65522]: _type = "Task" [ 640.089437] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.089693] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3f3e3cf3-e53c-4597-b660-7d019d158680 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.101047] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113543, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.102504] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 640.102504] env[65522]: value = "task-5113544" [ 640.102504] env[65522]: _type = "Task" [ 640.102504] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.114067] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113544, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.126584] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113540, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.836435} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.126859] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] bc40780e-43e2-4058-ab6f-d0d65e839128/bc40780e-43e2-4058-ab6f-d0d65e839128.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 640.127290] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 640.127598] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-37bead93-3d81-40fe-af1d-fb7e93c0a766 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.134842] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 640.134842] env[65522]: value = "task-5113545" [ 640.134842] env[65522]: _type = "Task" [ 640.134842] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.145142] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113545, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.157591] env[65522]: DEBUG nova.compute.manager [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 640.163817] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113541, 'name': CreateVM_Task, 'duration_secs': 0.537219} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.165897] env[65522]: WARNING neutronclient.v2_0.client [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.166411] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.166854] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.175951] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 640.177158] env[65522]: WARNING neutronclient.v2_0.client [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.177535] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.177761] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 640.178115] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 640.178580] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-412e4c86-84ad-4601-a4e1-8e34f92778cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.186387] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Waiting for the task: (returnval){ [ 640.186387] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52aa8d61-e997-57db-a33e-f3c80158092b" [ 640.186387] env[65522]: _type = "Task" [ 640.186387] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.192802] env[65522]: DEBUG nova.virt.hardware [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 640.193074] env[65522]: DEBUG nova.virt.hardware [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 640.193233] env[65522]: DEBUG nova.virt.hardware [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 640.193413] env[65522]: DEBUG nova.virt.hardware [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 640.193554] env[65522]: DEBUG nova.virt.hardware [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 640.193731] env[65522]: DEBUG nova.virt.hardware [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 640.193731] env[65522]: DEBUG nova.virt.hardware [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 640.194533] env[65522]: DEBUG nova.virt.hardware [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 640.194730] env[65522]: DEBUG nova.virt.hardware [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 640.194905] env[65522]: DEBUG nova.virt.hardware [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 640.195112] env[65522]: DEBUG nova.virt.hardware [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 640.196173] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70f3856-4e82-4c02-b57e-d88e6761a604 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.202375] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52aa8d61-e997-57db-a33e-f3c80158092b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.208679] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de85fe2a-083c-4bc7-90d0-47ce95b8df9d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.460025] env[65522]: DEBUG nova.compute.utils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 640.460025] env[65522]: DEBUG nova.compute.manager [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 640.460025] env[65522]: DEBUG nova.network.neutron [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 640.460651] env[65522]: WARNING neutronclient.v2_0.client [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.461173] env[65522]: WARNING neutronclient.v2_0.client [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.464022] env[65522]: WARNING openstack [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.464022] env[65522]: WARNING openstack [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.548267] env[65522]: DEBUG nova.network.neutron [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance_info_cache with network_info: [{"id": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "address": "fa:16:3e:69:05:d7", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf16c93a-75", "ovs_interfaceid": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 640.613406] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113543, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.620991] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113544, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.645638] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113545, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066767} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.646273] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 640.648323] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6a54e9-f073-47f7-ae2f-080d36f7dc5d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.679643] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Reconfiguring VM instance instance-00000008 to attach disk [datastore2] bc40780e-43e2-4058-ab6f-d0d65e839128/bc40780e-43e2-4058-ab6f-d0d65e839128.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 640.679643] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86d9d472-bff5-450b-a20b-da8dfc13a9d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.709900] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52aa8d61-e997-57db-a33e-f3c80158092b, 'name': SearchDatastore_Task, 'duration_secs': 0.047771} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.711710] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 640.712107] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 640.712324] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.712495] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 640.712645] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 640.712985] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 640.712985] env[65522]: value = "task-5113546" [ 640.712985] env[65522]: _type = "Task" [ 640.712985] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.713198] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-93b52eb2-7836-412d-8824-a1cd1d05231b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.727736] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113546, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.733269] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 640.733269] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 640.734162] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ca1e0f5-c3d4-496a-b9e3-bf9faeabdfd5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.741920] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Waiting for the task: (returnval){ [ 640.741920] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528572e1-f5d4-f12e-bba5-4a315dc12b6b" [ 640.741920] env[65522]: _type = "Task" [ 640.741920] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.752682] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528572e1-f5d4-f12e-bba5-4a315dc12b6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.760858] env[65522]: DEBUG nova.compute.manager [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Received event network-changed {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 640.761054] env[65522]: DEBUG nova.compute.manager [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Refreshing instance network info cache due to event network-changed. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 640.761275] env[65522]: DEBUG oslo_concurrency.lockutils [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] Acquiring lock "refresh_cache-dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.761409] env[65522]: DEBUG oslo_concurrency.lockutils [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] Acquired lock "refresh_cache-dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 640.761558] env[65522]: DEBUG nova.network.neutron [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 640.978379] env[65522]: DEBUG nova.compute.manager [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 641.053369] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.053840] env[65522]: DEBUG nova.compute.manager [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Instance network_info: |[{"id": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "address": "fa:16:3e:69:05:d7", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf16c93a-75", "ovs_interfaceid": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 641.054716] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:69:05:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4307c18-b235-43cd-bcd5-e226012d8ee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cf16c93a-755c-435b-8c86-833d8ea8afbf', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 641.064858] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Creating folder: Project (fa11b46d9fe144f391233e6eb9c819d7). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 641.068297] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dbbf9987-5939-4171-98ec-15316376fb77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.086222] env[65522]: DEBUG nova.policy [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '122b69071b0e4eb982aa687df5422092', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7340f053e9784f30be39030d99fe698f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 641.095823] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Created folder: Project (fa11b46d9fe144f391233e6eb9c819d7) in parent group-v994660. [ 641.096111] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Creating folder: Instances. Parent ref: group-v994688. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 641.096613] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53446633-18fc-4b79-aaf4-a6878de60d07 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.114918] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113543, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.117141] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Created folder: Instances in parent group-v994688. [ 641.117405] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 641.117604] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 641.117908] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3560b8e6-c289-4867-8258-0cef65282770 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.142707] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113544, 'name': Rename_Task, 'duration_secs': 0.529855} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.147581] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 641.148159] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf770e4a-ef93-477c-bc59-8ea71278104c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.153968] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 641.153968] env[65522]: value = "task-5113549" [ 641.153968] env[65522]: _type = "Task" [ 641.153968] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.163299] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 641.163299] env[65522]: value = "task-5113550" [ 641.163299] env[65522]: _type = "Task" [ 641.163299] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.172682] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113549, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.181894] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113550, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.226608] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113546, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.252800] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8769ab7e-cf1a-4a4f-9f79-1c6831a96e0f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.261081] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528572e1-f5d4-f12e-bba5-4a315dc12b6b, 'name': SearchDatastore_Task, 'duration_secs': 0.043641} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.263037] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e72ef256-d1c7-4317-8b76-ac55c31c1e61 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.268930] env[65522]: WARNING neutronclient.v2_0.client [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.275879] env[65522]: WARNING openstack [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.275879] env[65522]: WARNING openstack [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.391220] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4568ef6-3613-4a0e-987a-294ef3584178 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.424014] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Waiting for the task: (returnval){ [ 641.424014] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d0050d-c25d-fc8e-8e9d-a07c4a0f967b" [ 641.424014] env[65522]: _type = "Task" [ 641.424014] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.424014] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467fc415-f6c4-495c-a845-93afb0db3f3a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.440502] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d0050d-c25d-fc8e-8e9d-a07c4a0f967b, 'name': SearchDatastore_Task, 'duration_secs': 0.013435} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.440923] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.441278] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] a09c12be-0aaa-4934-a1b3-5af29feba8a6/a09c12be-0aaa-4934-a1b3-5af29feba8a6.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 641.442658] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db142720-41ad-4379-a433-5b6b207356ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.447267] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc4072dc-36bb-4ec1-83bf-b5551632bdaf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.465435] env[65522]: DEBUG nova.compute.provider_tree [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.468839] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Waiting for the task: (returnval){ [ 641.468839] env[65522]: value = "task-5113551" [ 641.468839] env[65522]: _type = "Task" [ 641.468839] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.479192] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113551, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.493777] env[65522]: DEBUG nova.network.neutron [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Updated VIF entry in instance network info cache for port dcf081b2-7a3e-4efd-84fa-bbf035dfd690. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 641.494362] env[65522]: DEBUG nova.network.neutron [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Updating instance_info_cache with network_info: [{"id": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "address": "fa:16:3e:9b:22:d2", "network": {"id": "b1e15367-b861-4a6c-8f94-921ed2b0a66f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1304672897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a47babe780ec4ae7b6c40f4531f09a38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "79c2e589-f55b-4843-8d99-2e565be16706", "external-id": "nsx-vlan-transportzone-858", "segmentation_id": 858, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcf081b2-7a", "ovs_interfaceid": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 641.504226] env[65522]: DEBUG nova.network.neutron [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Successfully updated port: 366aac9d-7922-4057-b141-5f4d54f4e2ab {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 641.554874] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.555303] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.610993] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113543, 'name': ReconfigVM_Task, 'duration_secs': 1.12037} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.611400] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Reconfigured VM instance instance-00000006 to attach disk [datastore2] 1d3a9ada-df46-46af-a7f1-a84c251af4ac/1d3a9ada-df46-46af-a7f1-a84c251af4ac.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 641.612135] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1ba9ee34-417c-4dbf-b5fe-a7a036a1276d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.621338] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 641.621338] env[65522]: value = "task-5113552" [ 641.621338] env[65522]: _type = "Task" [ 641.621338] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.632595] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113552, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.666576] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113549, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.677613] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113550, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.730445] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113546, 'name': ReconfigVM_Task, 'duration_secs': 0.761676} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.731228] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Reconfigured VM instance instance-00000008 to attach disk [datastore2] bc40780e-43e2-4058-ab6f-d0d65e839128/bc40780e-43e2-4058-ab6f-d0d65e839128.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 641.732192] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-04840c0e-ee6a-40ef-be3b-a64255f7b986 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.743524] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 641.743524] env[65522]: value = "task-5113553" [ 641.743524] env[65522]: _type = "Task" [ 641.743524] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.761666] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113553, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.972982] env[65522]: DEBUG nova.scheduler.client.report [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 641.990230] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113551, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.992620] env[65522]: WARNING neutronclient.v2_0.client [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.992620] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.992620] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.004461] env[65522]: DEBUG nova.compute.manager [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 642.006661] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Releasing lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.006998] env[65522]: DEBUG nova.compute.manager [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Received event network-vif-plugged-3896379b-7eb4-4dc5-819d-f68421a44dab {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 642.007309] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Acquiring lock "1d3a9ada-df46-46af-a7f1-a84c251af4ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.007462] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Lock "1d3a9ada-df46-46af-a7f1-a84c251af4ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.007661] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Lock "1d3a9ada-df46-46af-a7f1-a84c251af4ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.007896] env[65522]: DEBUG nova.compute.manager [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] No waiting events found dispatching network-vif-plugged-3896379b-7eb4-4dc5-819d-f68421a44dab {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 642.008169] env[65522]: WARNING nova.compute.manager [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Received unexpected event network-vif-plugged-3896379b-7eb4-4dc5-819d-f68421a44dab for instance with vm_state building and task_state spawning. [ 642.008546] env[65522]: DEBUG nova.compute.manager [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Received event network-changed-3896379b-7eb4-4dc5-819d-f68421a44dab {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 642.008713] env[65522]: DEBUG nova.compute.manager [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Refreshing instance network info cache due to event network-changed-3896379b-7eb4-4dc5-819d-f68421a44dab. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 642.008939] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Acquiring lock "refresh_cache-1d3a9ada-df46-46af-a7f1-a84c251af4ac" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.011019] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Acquired lock "refresh_cache-1d3a9ada-df46-46af-a7f1-a84c251af4ac" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.011019] env[65522]: DEBUG nova.network.neutron [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Refreshing network info cache for port 3896379b-7eb4-4dc5-819d-f68421a44dab {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 642.011752] env[65522]: DEBUG nova.network.neutron [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Successfully created port: ef361904-5296-49e1-aeb2-182cd4adb91b {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 642.019817] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Acquiring lock "refresh_cache-3f749f68-879f-44f3-8477-7ecea947078c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.020039] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Acquired lock "refresh_cache-3f749f68-879f-44f3-8477-7ecea947078c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.021231] env[65522]: DEBUG nova.network.neutron [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 642.024484] env[65522]: WARNING openstack [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.024484] env[65522]: WARNING openstack [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.033974] env[65522]: WARNING openstack [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.034306] env[65522]: WARNING openstack [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.051961] env[65522]: DEBUG nova.virt.hardware [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 642.052290] env[65522]: DEBUG nova.virt.hardware [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 642.052463] env[65522]: DEBUG nova.virt.hardware [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 642.053035] env[65522]: DEBUG nova.virt.hardware [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 642.053035] env[65522]: DEBUG nova.virt.hardware [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 642.053168] env[65522]: DEBUG nova.virt.hardware [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 642.053381] env[65522]: DEBUG nova.virt.hardware [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 642.053539] env[65522]: DEBUG nova.virt.hardware [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 642.054397] env[65522]: DEBUG nova.virt.hardware [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 642.054397] env[65522]: DEBUG nova.virt.hardware [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 642.054397] env[65522]: DEBUG nova.virt.hardware [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 642.055253] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d19387d0-4153-4ade-ad99-22b88219c7e9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.064412] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae712c1-6ac0-4756-8c55-51c83f203a83 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.138538] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113552, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.153066] env[65522]: DEBUG nova.network.neutron [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 642.166020] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113549, 'name': CreateVM_Task, 'duration_secs': 0.571352} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.166625] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 642.170579] env[65522]: WARNING neutronclient.v2_0.client [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 642.171140] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.171471] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.171880] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 642.172504] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1edf3155-a691-497a-bf51-0a96f66960ae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.180875] env[65522]: DEBUG oslo_vmware.api [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113550, 'name': PowerOnVM_Task, 'duration_secs': 0.777507} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.182688] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 642.182995] env[65522]: INFO nova.compute.manager [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Took 15.00 seconds to spawn the instance on the hypervisor. [ 642.183209] env[65522]: DEBUG nova.compute.manager [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 642.183545] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 642.183545] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5221aa34-9ff4-5a45-f6ea-8f69401213d6" [ 642.183545] env[65522]: _type = "Task" [ 642.183545] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.184270] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31fdd0bc-2cb0-41b7-8240-c16e4ff66e51 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.196424] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5221aa34-9ff4-5a45-f6ea-8f69401213d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.255600] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113553, 'name': Rename_Task, 'duration_secs': 0.260855} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.256034] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 642.258017] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dcb42acb-066b-4055-98a9-7ffaffebad4a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.265058] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 642.265058] env[65522]: value = "task-5113554" [ 642.265058] env[65522]: _type = "Task" [ 642.265058] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.274971] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113554, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.303908] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Updated VIF entry in instance network info cache for port 43c29c73-562e-48ba-886e-8e42d9c04074. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 642.304845] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Updating instance_info_cache with network_info: [{"id": "43c29c73-562e-48ba-886e-8e42d9c04074", "address": "fa:16:3e:7a:51:6e", "network": {"id": "2f4931f5-90c6-44a9-922e-e942d59ba127", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1285529142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f453b4359a94e54aca88fa2b9459476", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9a1e09ef-7c9c-45d9-9bf4-55b913524948", "external-id": "nsx-vlan-transportzone-466", "segmentation_id": 466, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c29c73-56", "ovs_interfaceid": "43c29c73-562e-48ba-886e-8e42d9c04074", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 642.389177] env[65522]: WARNING neutronclient.v2_0.client [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 642.389835] env[65522]: WARNING openstack [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.390282] env[65522]: WARNING openstack [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.408141] env[65522]: WARNING openstack [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.408613] env[65522]: WARNING openstack [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.471627] env[65522]: DEBUG nova.compute.manager [req-42e5178b-cae0-48f6-9e19-ec167f19ea22 req-e45b7f94-0aca-4847-8d8f-9234a9102155 service nova] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Received event network-vif-plugged-cf16c93a-755c-435b-8c86-833d8ea8afbf {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 642.472736] env[65522]: DEBUG oslo_concurrency.lockutils [req-42e5178b-cae0-48f6-9e19-ec167f19ea22 req-e45b7f94-0aca-4847-8d8f-9234a9102155 service nova] Acquiring lock "7c6f8218-602d-44f3-8012-de5a96972785-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.472736] env[65522]: DEBUG oslo_concurrency.lockutils [req-42e5178b-cae0-48f6-9e19-ec167f19ea22 req-e45b7f94-0aca-4847-8d8f-9234a9102155 service nova] Lock "7c6f8218-602d-44f3-8012-de5a96972785-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.472736] env[65522]: DEBUG oslo_concurrency.lockutils [req-42e5178b-cae0-48f6-9e19-ec167f19ea22 req-e45b7f94-0aca-4847-8d8f-9234a9102155 service nova] Lock "7c6f8218-602d-44f3-8012-de5a96972785-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.472736] env[65522]: DEBUG nova.compute.manager [req-42e5178b-cae0-48f6-9e19-ec167f19ea22 req-e45b7f94-0aca-4847-8d8f-9234a9102155 service nova] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] No waiting events found dispatching network-vif-plugged-cf16c93a-755c-435b-8c86-833d8ea8afbf {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 642.472736] env[65522]: WARNING nova.compute.manager [req-42e5178b-cae0-48f6-9e19-ec167f19ea22 req-e45b7f94-0aca-4847-8d8f-9234a9102155 service nova] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Received unexpected event network-vif-plugged-cf16c93a-755c-435b-8c86-833d8ea8afbf for instance with vm_state building and task_state spawning. [ 642.486427] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.488368] env[65522]: DEBUG nova.compute.manager [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 642.508482] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113551, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681533} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.508482] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] a09c12be-0aaa-4934-a1b3-5af29feba8a6/a09c12be-0aaa-4934-a1b3-5af29feba8a6.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 642.508482] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 642.508482] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb28a5de-9f92-4175-acf8-687156c4e4d8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.518630] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Waiting for the task: (returnval){ [ 642.518630] env[65522]: value = "task-5113555" [ 642.518630] env[65522]: _type = "Task" [ 642.518630] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.526652] env[65522]: WARNING neutronclient.v2_0.client [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 642.527321] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.527657] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.545954] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113555, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.641240] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113552, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.716769] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5221aa34-9ff4-5a45-f6ea-8f69401213d6, 'name': SearchDatastore_Task, 'duration_secs': 0.054712} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.717066] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.717312] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 642.717811] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.717811] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.718447] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 642.718447] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-658db9cd-8b2e-4b7c-8a06-681075653cff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.721337] env[65522]: INFO nova.compute.manager [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Took 21.58 seconds to build instance. [ 642.735169] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 642.735380] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 642.738136] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc1e8f83-2850-4285-82e2-c1650988258c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.751431] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 642.751431] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520a2e25-6196-4817-a86e-2ddb2442882e" [ 642.751431] env[65522]: _type = "Task" [ 642.751431] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.760326] env[65522]: DEBUG nova.network.neutron [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Updating instance_info_cache with network_info: [{"id": "e6e22697-3ece-4047-b378-b47c9c5de778", "address": "fa:16:3e:1a:4e:a7", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.26", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape6e22697-3e", "ovs_interfaceid": "e6e22697-3ece-4047-b378-b47c9c5de778", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 642.770342] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520a2e25-6196-4817-a86e-2ddb2442882e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.781832] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113554, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.809350] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Releasing lock "refresh_cache-fbb035b5-4562-44e3-9b95-8cf85299bb43" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.809628] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Received event network-vif-plugged-d24163e8-4f08-4908-9307-c4edb0d4ffb7 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 642.809833] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Acquiring lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.810216] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.810216] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.810410] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] No waiting events found dispatching network-vif-plugged-d24163e8-4f08-4908-9307-c4edb0d4ffb7 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 642.810593] env[65522]: WARNING nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Received unexpected event network-vif-plugged-d24163e8-4f08-4908-9307-c4edb0d4ffb7 for instance with vm_state building and task_state spawning. [ 642.810727] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Received event network-changed-d24163e8-4f08-4908-9307-c4edb0d4ffb7 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 642.811291] env[65522]: DEBUG nova.compute.manager [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Refreshing instance network info cache due to event network-changed-d24163e8-4f08-4908-9307-c4edb0d4ffb7. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 642.811291] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Acquiring lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.811291] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Acquired lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.811474] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Refreshing network info cache for port d24163e8-4f08-4908-9307-c4edb0d4ffb7 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 642.853485] env[65522]: WARNING neutronclient.v2_0.client [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 642.853485] env[65522]: WARNING openstack [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.853485] env[65522]: WARNING openstack [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.996910] env[65522]: DEBUG nova.compute.utils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 643.000757] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.001345] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.009726] env[65522]: DEBUG nova.compute.manager [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 643.010104] env[65522]: DEBUG nova.network.neutron [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 643.010414] env[65522]: WARNING neutronclient.v2_0.client [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.010753] env[65522]: WARNING neutronclient.v2_0.client [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.011417] env[65522]: WARNING openstack [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.011774] env[65522]: WARNING openstack [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.031701] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113555, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.13609} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.032074] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 643.032989] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153759e6-9d89-49f1-9c64-09affd41abd4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.060970] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Reconfiguring VM instance instance-00000009 to attach disk [datastore2] a09c12be-0aaa-4934-a1b3-5af29feba8a6/a09c12be-0aaa-4934-a1b3-5af29feba8a6.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 643.063279] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-525c2a73-31a1-4ffa-a668-73a21940ecc6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.087593] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Waiting for the task: (returnval){ [ 643.087593] env[65522]: value = "task-5113556" [ 643.087593] env[65522]: _type = "Task" [ 643.087593] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.100546] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113556, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.137129] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113552, 'name': Rename_Task, 'duration_secs': 1.231027} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.137525] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 643.137962] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-072d9b16-6a84-456c-a083-a797ddea322e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.147284] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 643.147284] env[65522]: value = "task-5113557" [ 643.147284] env[65522]: _type = "Task" [ 643.147284] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.158485] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113557, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.224071] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34694d97-3021-47d8-87fa-22353fa14dae tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Lock "6abf58ca-ad6b-4fe8-844a-ca33e62a844d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.094s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 643.263913] env[65522]: DEBUG oslo_concurrency.lockutils [None req-405534ba-21d2-4706-80c7-9a71503d6159 tempest-ServerExternalEventsTest-1082039177 tempest-ServerExternalEventsTest-1082039177-project] Releasing lock "refresh_cache-dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 643.264362] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520a2e25-6196-4817-a86e-2ddb2442882e, 'name': SearchDatastore_Task, 'duration_secs': 0.023177} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.265218] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2a9b447-ec76-4129-859d-d02cd1afc162 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.275803] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 643.275803] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527eb6ef-1653-b32e-3820-dce8604cb47f" [ 643.275803] env[65522]: _type = "Task" [ 643.275803] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.283957] env[65522]: DEBUG oslo_vmware.api [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113554, 'name': PowerOnVM_Task, 'duration_secs': 0.943132} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.284722] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 643.284998] env[65522]: INFO nova.compute.manager [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Took 11.34 seconds to spawn the instance on the hypervisor. [ 643.285241] env[65522]: DEBUG nova.compute.manager [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 643.286154] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96cfce5f-be2a-45c7-8424-10594b536194 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.293293] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527eb6ef-1653-b32e-3820-dce8604cb47f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.310261] env[65522]: DEBUG nova.network.neutron [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Updating instance_info_cache with network_info: [{"id": "366aac9d-7922-4057-b141-5f4d54f4e2ab", "address": "fa:16:3e:66:0d:26", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.217", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap366aac9d-79", "ovs_interfaceid": "366aac9d-7922-4057-b141-5f4d54f4e2ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 643.321738] env[65522]: WARNING neutronclient.v2_0.client [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.322492] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.322838] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.500496] env[65522]: DEBUG nova.compute.manager [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 643.597304] env[65522]: DEBUG nova.policy [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '792013a276474561bc625ac0e00cdb82', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2bd38f300424db78568e0b8d9982f27', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 643.603646] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113556, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.661823] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113557, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.748536] env[65522]: DEBUG nova.network.neutron [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Successfully updated port: ef361904-5296-49e1-aeb2-182cd4adb91b {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 643.788510] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527eb6ef-1653-b32e-3820-dce8604cb47f, 'name': SearchDatastore_Task, 'duration_secs': 0.020087} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.788624] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 643.788886] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 7c6f8218-602d-44f3-8012-de5a96972785/7c6f8218-602d-44f3-8012-de5a96972785.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 643.789173] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-00a38b95-a0e0-45ae-872b-55efc3526bcb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.798177] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 643.798177] env[65522]: value = "task-5113558" [ 643.798177] env[65522]: _type = "Task" [ 643.798177] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.815042] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113558, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.815042] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Releasing lock "refresh_cache-3f749f68-879f-44f3-8477-7ecea947078c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 643.815242] env[65522]: DEBUG nova.compute.manager [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Instance network_info: |[{"id": "366aac9d-7922-4057-b141-5f4d54f4e2ab", "address": "fa:16:3e:66:0d:26", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.217", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap366aac9d-79", "ovs_interfaceid": "366aac9d-7922-4057-b141-5f4d54f4e2ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 643.815353] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:0d:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '366aac9d-7922-4057-b141-5f4d54f4e2ab', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 643.823158] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Creating folder: Project (e099bf52db6241d78a15fed1c5bbfbc8). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 643.823766] env[65522]: INFO nova.compute.manager [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Took 19.46 seconds to build instance. [ 643.825312] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cdb5ae5d-7cdf-4a23-91b3-f45bf43b86ce {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.838359] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Created folder: Project (e099bf52db6241d78a15fed1c5bbfbc8) in parent group-v994660. [ 643.838583] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Creating folder: Instances. Parent ref: group-v994691. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 643.838898] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5076c123-5363-45c1-88fd-3664da2e8b8b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.852158] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Created folder: Instances in parent group-v994691. [ 643.852158] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 643.852158] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 643.852158] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04892262-d256-45eb-99b8-ca4d374f2c03 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.877382] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 643.877382] env[65522]: value = "task-5113561" [ 643.877382] env[65522]: _type = "Task" [ 643.877382] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.891836] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113561, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.058425] env[65522]: WARNING neutronclient.v2_0.client [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 644.059336] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.059718] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.100801] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113556, 'name': ReconfigVM_Task, 'duration_secs': 0.643627} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.101186] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Reconfigured VM instance instance-00000009 to attach disk [datastore2] a09c12be-0aaa-4934-a1b3-5af29feba8a6/a09c12be-0aaa-4934-a1b3-5af29feba8a6.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 644.101914] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-48dc4a92-e509-422f-ac02-3b210a37106a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.110856] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Waiting for the task: (returnval){ [ 644.110856] env[65522]: value = "task-5113562" [ 644.110856] env[65522]: _type = "Task" [ 644.110856] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.121989] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113562, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.158347] env[65522]: DEBUG oslo_vmware.api [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113557, 'name': PowerOnVM_Task, 'duration_secs': 0.815945} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.158646] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 644.158865] env[65522]: INFO nova.compute.manager [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Took 14.64 seconds to spawn the instance on the hypervisor. [ 644.159077] env[65522]: DEBUG nova.compute.manager [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 644.159989] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada857a6-bfa9-4b94-955b-4af9ede71c3b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.254608] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Acquiring lock "refresh_cache-2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.254802] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Acquired lock "refresh_cache-2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 644.254979] env[65522]: DEBUG nova.network.neutron [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 644.258167] env[65522]: DEBUG nova.network.neutron [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Successfully created port: 2eec2687-b0a0-442d-8306-27ab523bf504 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 644.312720] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113558, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.328379] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f550d86-1a5a-4de8-9faa-6ecb54fe4d57 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lock "bc40780e-43e2-4058-ab6f-d0d65e839128" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.981s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 644.389627] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113561, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.415104] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.415650] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.511138] env[65522]: DEBUG nova.compute.manager [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 644.547785] env[65522]: DEBUG nova.virt.hardware [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 644.548376] env[65522]: DEBUG nova.virt.hardware [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 644.548644] env[65522]: DEBUG nova.virt.hardware [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 644.548925] env[65522]: DEBUG nova.virt.hardware [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 644.549155] env[65522]: DEBUG nova.virt.hardware [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 644.549374] env[65522]: DEBUG nova.virt.hardware [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 644.549692] env[65522]: DEBUG nova.virt.hardware [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 644.549943] env[65522]: DEBUG nova.virt.hardware [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 644.550245] env[65522]: DEBUG nova.virt.hardware [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 644.550495] env[65522]: DEBUG nova.virt.hardware [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 644.551197] env[65522]: DEBUG nova.virt.hardware [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 644.553560] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f37b1dd1-5f79-4256-92bc-cfd38ca7ee0f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.566252] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b174a85-26d8-4c11-8145-03e9e9b69ffb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.623968] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113562, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.681083] env[65522]: INFO nova.compute.manager [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Took 23.49 seconds to build instance. [ 644.758640] env[65522]: WARNING openstack [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.759168] env[65522]: WARNING openstack [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.809598] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113558, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.845312} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.809862] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 7c6f8218-602d-44f3-8012-de5a96972785/7c6f8218-602d-44f3-8012-de5a96972785.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 644.810344] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 644.810474] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc8cb590-526e-4948-a106-2139add776e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.817913] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 644.817913] env[65522]: value = "task-5113563" [ 644.817913] env[65522]: _type = "Task" [ 644.817913] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.828316] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113563, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.888887] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113561, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.121995] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113562, 'name': Rename_Task, 'duration_secs': 0.975673} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.122282] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 645.122572] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5327ac9-f347-40d1-9860-14a41610e5b2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.130852] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Waiting for the task: (returnval){ [ 645.130852] env[65522]: value = "task-5113564" [ 645.130852] env[65522]: _type = "Task" [ 645.130852] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.142126] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113564, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.183920] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c433fb77-6f24-4c4c-b9da-ee5367e17525 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lock "1d3a9ada-df46-46af-a7f1-a84c251af4ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 645.197640] env[65522]: DEBUG nova.network.neutron [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Updated VIF entry in instance network info cache for port 3896379b-7eb4-4dc5-819d-f68421a44dab. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 645.198150] env[65522]: DEBUG nova.network.neutron [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Updating instance_info_cache with network_info: [{"id": "3896379b-7eb4-4dc5-819d-f68421a44dab", "address": "fa:16:3e:c8:84:60", "network": {"id": "2dbba2da-d7e7-45b3-953b-aeb93ddb6177", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-737085066-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "345c026cba1d40c59cab5aa5b477b833", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3896379b-7e", "ovs_interfaceid": "3896379b-7eb4-4dc5-819d-f68421a44dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 645.331667] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113563, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07122} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.331878] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 645.332832] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36d6d9f-fc25-4987-a86a-21505fa7e3b9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.358363] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Reconfiguring VM instance instance-0000000a to attach disk [datastore2] 7c6f8218-602d-44f3-8012-de5a96972785/7c6f8218-602d-44f3-8012-de5a96972785.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 645.358763] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1511a5fa-022b-4ec7-9a29-133cc13ed7cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.383833] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 645.383833] env[65522]: value = "task-5113565" [ 645.383833] env[65522]: _type = "Task" [ 645.383833] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.387911] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113561, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.398536] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113565, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.644055] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113564, 'name': PowerOnVM_Task} progress is 1%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.701137] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Releasing lock "refresh_cache-1d3a9ada-df46-46af-a7f1-a84c251af4ac" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 645.701562] env[65522]: DEBUG nova.compute.manager [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Received event network-vif-plugged-9902660c-976f-4cee-8338-74f077c8f7b3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 645.701873] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Acquiring lock "bc40780e-43e2-4058-ab6f-d0d65e839128-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 645.702404] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Lock "bc40780e-43e2-4058-ab6f-d0d65e839128-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 645.702552] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Lock "bc40780e-43e2-4058-ab6f-d0d65e839128-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 645.703110] env[65522]: DEBUG nova.compute.manager [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] No waiting events found dispatching network-vif-plugged-9902660c-976f-4cee-8338-74f077c8f7b3 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 645.703355] env[65522]: WARNING nova.compute.manager [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Received unexpected event network-vif-plugged-9902660c-976f-4cee-8338-74f077c8f7b3 for instance with vm_state building and task_state spawning. [ 645.703530] env[65522]: DEBUG nova.compute.manager [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Received event network-changed-9902660c-976f-4cee-8338-74f077c8f7b3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 645.703690] env[65522]: DEBUG nova.compute.manager [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Refreshing instance network info cache due to event network-changed-9902660c-976f-4cee-8338-74f077c8f7b3. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 645.703888] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Acquiring lock "refresh_cache-bc40780e-43e2-4058-ab6f-d0d65e839128" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.704473] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Acquired lock "refresh_cache-bc40780e-43e2-4058-ab6f-d0d65e839128" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 645.704473] env[65522]: DEBUG nova.network.neutron [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Refreshing network info cache for port 9902660c-976f-4cee-8338-74f077c8f7b3 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 645.889832] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113561, 'name': CreateVM_Task, 'duration_secs': 1.606038} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.893155] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 645.893748] env[65522]: WARNING neutronclient.v2_0.client [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.894263] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.894333] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 645.894657] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 645.895506] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0663ddad-2657-4474-875c-8f2372856494 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.902478] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113565, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.904232] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Waiting for the task: (returnval){ [ 645.904232] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a2897b-14b6-b7a3-f1b6-5305832a808a" [ 645.904232] env[65522]: _type = "Task" [ 645.904232] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.915305] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a2897b-14b6-b7a3-f1b6-5305832a808a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.927839] env[65522]: DEBUG nova.network.neutron [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 645.958125] env[65522]: WARNING neutronclient.v2_0.client [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.959110] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.959435] env[65522]: WARNING openstack [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.065728] env[65522]: WARNING openstack [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.066302] env[65522]: WARNING openstack [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.085789] env[65522]: DEBUG nova.compute.manager [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Received event network-vif-plugged-a002c980-b974-432b-aa4e-e8613c617e8e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 646.086012] env[65522]: DEBUG oslo_concurrency.lockutils [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] Acquiring lock "a09c12be-0aaa-4934-a1b3-5af29feba8a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.086998] env[65522]: DEBUG oslo_concurrency.lockutils [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] Lock "a09c12be-0aaa-4934-a1b3-5af29feba8a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.086998] env[65522]: DEBUG oslo_concurrency.lockutils [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] Lock "a09c12be-0aaa-4934-a1b3-5af29feba8a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.089210] env[65522]: DEBUG nova.compute.manager [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] No waiting events found dispatching network-vif-plugged-a002c980-b974-432b-aa4e-e8613c617e8e {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 646.089461] env[65522]: WARNING nova.compute.manager [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Received unexpected event network-vif-plugged-a002c980-b974-432b-aa4e-e8613c617e8e for instance with vm_state building and task_state spawning. [ 646.089690] env[65522]: DEBUG nova.compute.manager [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Received event network-changed-a002c980-b974-432b-aa4e-e8613c617e8e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 646.089860] env[65522]: DEBUG nova.compute.manager [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Refreshing instance network info cache due to event network-changed-a002c980-b974-432b-aa4e-e8613c617e8e. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 646.090091] env[65522]: DEBUG oslo_concurrency.lockutils [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] Acquiring lock "refresh_cache-a09c12be-0aaa-4934-a1b3-5af29feba8a6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.090227] env[65522]: DEBUG oslo_concurrency.lockutils [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] Acquired lock "refresh_cache-a09c12be-0aaa-4934-a1b3-5af29feba8a6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 646.090408] env[65522]: DEBUG nova.network.neutron [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Refreshing network info cache for port a002c980-b974-432b-aa4e-e8613c617e8e {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 646.095200] env[65522]: DEBUG nova.network.neutron [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Successfully updated port: 2eec2687-b0a0-442d-8306-27ab523bf504 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 646.145936] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113564, 'name': PowerOnVM_Task} progress is 64%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.175134] env[65522]: WARNING neutronclient.v2_0.client [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.175814] env[65522]: WARNING openstack [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.176165] env[65522]: WARNING openstack [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.207860] env[65522]: WARNING neutronclient.v2_0.client [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.208602] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.209635] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.314474] env[65522]: DEBUG nova.network.neutron [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Updating instance_info_cache with network_info: [{"id": "ef361904-5296-49e1-aeb2-182cd4adb91b", "address": "fa:16:3e:58:20:55", "network": {"id": "d13db148-3532-452c-8c04-2a9a32a8c145", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1279941651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7340f053e9784f30be39030d99fe698f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f969bd9-e040-4b9b-85b2-7c61231584ad", "external-id": "nsx-vlan-transportzone-995", "segmentation_id": 995, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef361904-52", "ovs_interfaceid": "ef361904-5296-49e1-aeb2-182cd4adb91b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 646.403037] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113565, 'name': ReconfigVM_Task, 'duration_secs': 0.600566} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.403559] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Reconfigured VM instance instance-0000000a to attach disk [datastore2] 7c6f8218-602d-44f3-8012-de5a96972785/7c6f8218-602d-44f3-8012-de5a96972785.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 646.407452] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-265fbacb-87ae-497d-add7-dede4a869154 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.420626] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a2897b-14b6-b7a3-f1b6-5305832a808a, 'name': SearchDatastore_Task, 'duration_secs': 0.014292} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.422981] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 646.424223] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 646.424975] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.425259] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 646.425576] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 646.426104] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 646.426104] env[65522]: value = "task-5113566" [ 646.426104] env[65522]: _type = "Task" [ 646.426104] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.426509] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d07df113-9641-4029-a821-91df8b017593 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.439729] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113566, 'name': Rename_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.444574] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 646.444574] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 646.444574] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c90b8d58-b01c-4553-9b82-aecf7bc041a6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.452922] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Waiting for the task: (returnval){ [ 646.452922] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ea0d94-3ff7-3a88-8d47-0f2aa919cd22" [ 646.452922] env[65522]: _type = "Task" [ 646.452922] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.466532] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ea0d94-3ff7-3a88-8d47-0f2aa919cd22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.497593] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.498075] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.548767] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Updated VIF entry in instance network info cache for port d24163e8-4f08-4908-9307-c4edb0d4ffb7. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 646.549374] env[65522]: DEBUG nova.network.neutron [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Updating instance_info_cache with network_info: [{"id": "d24163e8-4f08-4908-9307-c4edb0d4ffb7", "address": "fa:16:3e:e2:90:31", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd24163e8-4f", "ovs_interfaceid": "d24163e8-4f08-4908-9307-c4edb0d4ffb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 646.598810] env[65522]: WARNING neutronclient.v2_0.client [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.599513] env[65522]: WARNING openstack [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.599958] env[65522]: WARNING openstack [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.614830] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "refresh_cache-6c8165f5-6769-4954-b9ac-de8551954377" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.614830] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquired lock "refresh_cache-6c8165f5-6769-4954-b9ac-de8551954377" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 646.617347] env[65522]: DEBUG nova.network.neutron [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 646.644258] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113564, 'name': PowerOnVM_Task} progress is 82%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.694545] env[65522]: WARNING neutronclient.v2_0.client [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.695227] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.696126] env[65522]: WARNING openstack [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.722728] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Acquiring lock "fbb035b5-4562-44e3-9b95-8cf85299bb43" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.723106] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Lock "fbb035b5-4562-44e3-9b95-8cf85299bb43" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.723156] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Acquiring lock "fbb035b5-4562-44e3-9b95-8cf85299bb43-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.724129] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Lock "fbb035b5-4562-44e3-9b95-8cf85299bb43-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.724129] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Lock "fbb035b5-4562-44e3-9b95-8cf85299bb43-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.726349] env[65522]: INFO nova.compute.manager [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Terminating instance [ 646.762902] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Acquiring lock "dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.763141] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Lock "dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.763297] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Acquiring lock "dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.763464] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Lock "dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.763620] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Lock "dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.765912] env[65522]: INFO nova.compute.manager [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Terminating instance [ 646.818725] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Releasing lock "refresh_cache-2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 646.819153] env[65522]: DEBUG nova.compute.manager [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Instance network_info: |[{"id": "ef361904-5296-49e1-aeb2-182cd4adb91b", "address": "fa:16:3e:58:20:55", "network": {"id": "d13db148-3532-452c-8c04-2a9a32a8c145", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1279941651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7340f053e9784f30be39030d99fe698f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f969bd9-e040-4b9b-85b2-7c61231584ad", "external-id": "nsx-vlan-transportzone-995", "segmentation_id": 995, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef361904-52", "ovs_interfaceid": "ef361904-5296-49e1-aeb2-182cd4adb91b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 646.819675] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:58:20:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6f969bd9-e040-4b9b-85b2-7c61231584ad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef361904-5296-49e1-aeb2-182cd4adb91b', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 646.828018] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Creating folder: Project (7340f053e9784f30be39030d99fe698f). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 646.828741] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a5f453f8-71cf-4cde-b98d-e30a7ee4b50d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.840760] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Created folder: Project (7340f053e9784f30be39030d99fe698f) in parent group-v994660. [ 646.841084] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Creating folder: Instances. Parent ref: group-v994694. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 646.841352] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd853f4c-9481-4592-a57e-4d0925c93fdc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.851903] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Created folder: Instances in parent group-v994694. [ 646.852203] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 646.852418] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 646.852635] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-abeef3e9-d589-428a-986d-c74387fd0702 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.873794] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 646.873794] env[65522]: value = "task-5113569" [ 646.873794] env[65522]: _type = "Task" [ 646.873794] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.885850] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113569, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.925644] env[65522]: DEBUG nova.network.neutron [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Updated VIF entry in instance network info cache for port 9902660c-976f-4cee-8338-74f077c8f7b3. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 646.925644] env[65522]: DEBUG nova.network.neutron [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Updating instance_info_cache with network_info: [{"id": "9902660c-976f-4cee-8338-74f077c8f7b3", "address": "fa:16:3e:20:4e:b5", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.243", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9902660c-97", "ovs_interfaceid": "9902660c-976f-4cee-8338-74f077c8f7b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 646.940018] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113566, 'name': Rename_Task, 'duration_secs': 0.21548} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.941455] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 646.941455] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1c29ae6-0320-449e-bb8b-1b87efd6bb9b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.950466] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 646.950466] env[65522]: value = "task-5113570" [ 646.950466] env[65522]: _type = "Task" [ 646.950466] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.970506] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.974337] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ea0d94-3ff7-3a88-8d47-0f2aa919cd22, 'name': SearchDatastore_Task, 'duration_secs': 0.016048} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.975280] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a18f601-21ae-459f-8b14-55b67e4a2968 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.982519] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Waiting for the task: (returnval){ [ 646.982519] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cbba3b-5020-2d33-9c8f-116d6cffeaf8" [ 646.982519] env[65522]: _type = "Task" [ 646.982519] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.996834] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cbba3b-5020-2d33-9c8f-116d6cffeaf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.009875] env[65522]: WARNING openstack [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.010328] env[65522]: WARNING openstack [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.053061] env[65522]: DEBUG oslo_concurrency.lockutils [req-a51097ee-1449-41dd-8c42-b5b62ec50d4e req-0005656c-ac9c-4447-9930-722352ccccd4 service nova] Releasing lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 647.124697] env[65522]: WARNING openstack [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.125643] env[65522]: WARNING openstack [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.149945] env[65522]: DEBUG oslo_vmware.api [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113564, 'name': PowerOnVM_Task, 'duration_secs': 1.676547} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.150373] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 647.150673] env[65522]: INFO nova.compute.manager [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Took 12.78 seconds to spawn the instance on the hypervisor. [ 647.151064] env[65522]: DEBUG nova.compute.manager [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 647.152424] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ea7983-4f9b-4a0b-b064-7a409df46925 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.232807] env[65522]: DEBUG nova.compute.manager [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 647.233105] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 647.235715] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b801fa0a-7131-4ce2-9be4-d89ae379c98b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.252464] env[65522]: WARNING neutronclient.v2_0.client [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 647.253435] env[65522]: WARNING openstack [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.254729] env[65522]: WARNING openstack [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.272998] env[65522]: DEBUG nova.compute.manager [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 647.273343] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 647.273766] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 647.276445] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-341c86fa-148a-4e1b-a647-41e07d12ac9a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.280926] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db2154a8-4ba5-4695-82c5-2d1788f4ee0a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.288146] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 647.290171] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4451fa04-1c31-4b70-8c39-02b4a1236df2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.292216] env[65522]: DEBUG oslo_vmware.api [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Waiting for the task: (returnval){ [ 647.292216] env[65522]: value = "task-5113571" [ 647.292216] env[65522]: _type = "Task" [ 647.292216] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.299018] env[65522]: DEBUG oslo_vmware.api [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Waiting for the task: (returnval){ [ 647.299018] env[65522]: value = "task-5113572" [ 647.299018] env[65522]: _type = "Task" [ 647.299018] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.306962] env[65522]: DEBUG oslo_vmware.api [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113571, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.315452] env[65522]: DEBUG oslo_vmware.api [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113572, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.380356] env[65522]: DEBUG nova.network.neutron [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 647.388489] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113569, 'name': CreateVM_Task} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.388706] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 647.389190] env[65522]: WARNING neutronclient.v2_0.client [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 647.389555] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.389702] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 647.390034] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 647.390355] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9456eef2-e517-4cfa-bbc2-7c2380eca193 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.396420] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Waiting for the task: (returnval){ [ 647.396420] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52779727-7e76-b262-dc04-1cc3ea5d1465" [ 647.396420] env[65522]: _type = "Task" [ 647.396420] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.407595] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52779727-7e76-b262-dc04-1cc3ea5d1465, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.427520] env[65522]: DEBUG oslo_concurrency.lockutils [req-d24af95a-9784-4d00-bdec-9b3b8bca1a20 req-92dcc4c2-c798-40fc-b86e-8ae78a2122b1 service nova] Releasing lock "refresh_cache-bc40780e-43e2-4058-ab6f-d0d65e839128" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 647.462580] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113570, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.494533] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cbba3b-5020-2d33-9c8f-116d6cffeaf8, 'name': SearchDatastore_Task, 'duration_secs': 0.015197} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.494820] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 647.495216] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 3f749f68-879f-44f3-8477-7ecea947078c/3f749f68-879f-44f3-8477-7ecea947078c.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 647.495507] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0324d4cb-b2a4-4d64-90ab-55c98c76f52d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.505598] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Waiting for the task: (returnval){ [ 647.505598] env[65522]: value = "task-5113573" [ 647.505598] env[65522]: _type = "Task" [ 647.505598] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.517019] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113573, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.620386] env[65522]: WARNING openstack [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.620770] env[65522]: WARNING openstack [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.676731] env[65522]: INFO nova.compute.manager [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Took 23.16 seconds to build instance. [ 647.733359] env[65522]: DEBUG nova.network.neutron [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Updated VIF entry in instance network info cache for port a002c980-b974-432b-aa4e-e8613c617e8e. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 647.734016] env[65522]: DEBUG nova.network.neutron [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Updating instance_info_cache with network_info: [{"id": "a002c980-b974-432b-aa4e-e8613c617e8e", "address": "fa:16:3e:d4:83:9e", "network": {"id": "65a80e4b-308a-433d-8c2d-4f21879d8ac2", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1688898810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1927d35d9154557a33422045152daa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f847601f-7479-48eb-842f-41f94eea8537", "external-id": "nsx-vlan-transportzone-35", "segmentation_id": 35, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa002c980-b9", "ovs_interfaceid": "a002c980-b974-432b-aa4e-e8613c617e8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 647.802629] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "ad074365-1443-481d-990f-479c523219fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 647.802892] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "ad074365-1443-481d-990f-479c523219fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 647.815155] env[65522]: DEBUG oslo_vmware.api [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113571, 'name': PowerOffVM_Task, 'duration_secs': 0.245784} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.820360] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 647.820360] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 647.820979] env[65522]: DEBUG oslo_vmware.api [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113572, 'name': PowerOffVM_Task, 'duration_secs': 0.219959} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.821265] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff2c5e23-57fc-453a-ad5f-cd10757d15d5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.823231] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 647.823419] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 647.824239] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f95af0f-c337-44c6-b41d-2edb80457ae3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.913248] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52779727-7e76-b262-dc04-1cc3ea5d1465, 'name': SearchDatastore_Task, 'duration_secs': 0.014261} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.913248] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 647.913743] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 647.916621] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.916621] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 647.916621] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 647.916621] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b5ca5913-f5bd-4ebf-b64b-3fda26c08fcb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.927145] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 647.927493] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 647.928212] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Deleting the datastore file [datastore1] dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 647.929567] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e032c528-7e5d-4b56-a796-ee74e3469f1f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.931865] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 647.932460] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 647.932700] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Deleting the datastore file [datastore1] fbb035b5-4562-44e3-9b95-8cf85299bb43 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 647.934170] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0abef7a1-b8c0-4793-ad1a-e1ed5b1b91d9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.936886] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 647.937060] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 647.938322] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfb2e2eb-d32f-428a-ac0d-8cae158af36f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.947494] env[65522]: DEBUG oslo_vmware.api [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Waiting for the task: (returnval){ [ 647.947494] env[65522]: value = "task-5113576" [ 647.947494] env[65522]: _type = "Task" [ 647.947494] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.949849] env[65522]: DEBUG oslo_vmware.api [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Waiting for the task: (returnval){ [ 647.949849] env[65522]: value = "task-5113577" [ 647.949849] env[65522]: _type = "Task" [ 647.949849] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.950369] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Waiting for the task: (returnval){ [ 647.950369] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5299d56d-e25f-a201-bd70-78ea795a4889" [ 647.950369] env[65522]: _type = "Task" [ 647.950369] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.972202] env[65522]: INFO nova.compute.manager [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Rebuilding instance [ 647.981908] env[65522]: DEBUG oslo_vmware.api [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113576, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.992119] env[65522]: DEBUG oslo_vmware.api [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113577, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.992703] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5299d56d-e25f-a201-bd70-78ea795a4889, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.993204] env[65522]: DEBUG oslo_vmware.api [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113570, 'name': PowerOnVM_Task, 'duration_secs': 0.649696} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.995338] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 647.995539] env[65522]: INFO nova.compute.manager [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Took 11.17 seconds to spawn the instance on the hypervisor. [ 647.996113] env[65522]: DEBUG nova.compute.manager [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 647.998220] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2732e9bd-5ce7-4ddf-9d10-6e5b7378d6f5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.025567] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113573, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.034823] env[65522]: DEBUG nova.compute.manager [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 648.035822] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e76204-baa0-461f-804a-12a4a9f926fa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.074037] env[65522]: WARNING neutronclient.v2_0.client [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.074889] env[65522]: WARNING openstack [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.075262] env[65522]: WARNING openstack [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.181176] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d3a7a812-a797-4a64-9400-d72c20a7aa2f tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Lock "a09c12be-0aaa-4934-a1b3-5af29feba8a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.676s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 648.240707] env[65522]: DEBUG oslo_concurrency.lockutils [req-cbc33cca-3e7e-4095-8cdc-0e0d3c2236df req-8a93defe-7e0d-483d-ae25-f4e3a7eb62d1 service nova] Releasing lock "refresh_cache-a09c12be-0aaa-4934-a1b3-5af29feba8a6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 648.248074] env[65522]: DEBUG nova.network.neutron [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Updating instance_info_cache with network_info: [{"id": "2eec2687-b0a0-442d-8306-27ab523bf504", "address": "fa:16:3e:0c:f9:44", "network": {"id": "472285e2-f5f3-4616-92f7-4c0144d833d4", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-945513036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "e2bd38f300424db78568e0b8d9982f27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2eec2687-b0", "ovs_interfaceid": "2eec2687-b0a0-442d-8306-27ab523bf504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 648.263034] env[65522]: DEBUG nova.compute.manager [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Received event network-changed-cf16c93a-755c-435b-8c86-833d8ea8afbf {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 648.263457] env[65522]: DEBUG nova.compute.manager [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Refreshing instance network info cache due to event network-changed-cf16c93a-755c-435b-8c86-833d8ea8afbf. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 648.263457] env[65522]: DEBUG oslo_concurrency.lockutils [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] Acquiring lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.263626] env[65522]: DEBUG oslo_concurrency.lockutils [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] Acquired lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 648.264127] env[65522]: DEBUG nova.network.neutron [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Refreshing network info cache for port cf16c93a-755c-435b-8c86-833d8ea8afbf {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 648.309539] env[65522]: DEBUG nova.compute.manager [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 648.469615] env[65522]: DEBUG oslo_vmware.api [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Task: {'id': task-5113576, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.495457} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.476094] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 648.476359] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 648.476540] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 648.476725] env[65522]: INFO nova.compute.manager [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Took 1.20 seconds to destroy the instance on the hypervisor. [ 648.477029] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 648.478184] env[65522]: DEBUG oslo_vmware.api [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113577, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.478184] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5299d56d-e25f-a201-bd70-78ea795a4889, 'name': SearchDatastore_Task, 'duration_secs': 0.035249} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.478184] env[65522]: DEBUG nova.compute.manager [-] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 648.478184] env[65522]: DEBUG nova.network.neutron [-] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 648.478184] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.478613] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.478876] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.486638] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc00ea1b-c8f2-4837-a7f3-a04c66b06900 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.493639] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Waiting for the task: (returnval){ [ 648.493639] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e6e3d3-822b-518f-8296-60c8e74d1b72" [ 648.493639] env[65522]: _type = "Task" [ 648.493639] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.503751] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e6e3d3-822b-518f-8296-60c8e74d1b72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.522643] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113573, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.684709} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.525687] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 3f749f68-879f-44f3-8477-7ecea947078c/3f749f68-879f-44f3-8477-7ecea947078c.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 648.525687] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 648.525687] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e6ded566-0e3c-4b20-bcc9-bfe19e7049d1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.530380] env[65522]: INFO nova.compute.manager [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Took 22.81 seconds to build instance. [ 648.537545] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Waiting for the task: (returnval){ [ 648.537545] env[65522]: value = "task-5113578" [ 648.537545] env[65522]: _type = "Task" [ 648.537545] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.549167] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113578, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.577367] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.752108] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Releasing lock "refresh_cache-6c8165f5-6769-4954-b9ac-de8551954377" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 648.752108] env[65522]: DEBUG nova.compute.manager [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Instance network_info: |[{"id": "2eec2687-b0a0-442d-8306-27ab523bf504", "address": "fa:16:3e:0c:f9:44", "network": {"id": "472285e2-f5f3-4616-92f7-4c0144d833d4", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-945513036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "e2bd38f300424db78568e0b8d9982f27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2eec2687-b0", "ovs_interfaceid": "2eec2687-b0a0-442d-8306-27ab523bf504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 648.752481] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:f9:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2eec2687-b0a0-442d-8306-27ab523bf504', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 648.759764] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Creating folder: Project (e2bd38f300424db78568e0b8d9982f27). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 648.760117] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-de67bde6-8e2e-44a6-990c-4dc7205e987d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.771967] env[65522]: WARNING neutronclient.v2_0.client [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.772522] env[65522]: WARNING openstack [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.772886] env[65522]: WARNING openstack [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.780191] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Created folder: Project (e2bd38f300424db78568e0b8d9982f27) in parent group-v994660. [ 648.780319] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Creating folder: Instances. Parent ref: group-v994697. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 648.782991] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6126e8c5-26be-4428-b817-ac0a2035f9e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.794062] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Created folder: Instances in parent group-v994697. [ 648.794062] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 648.794062] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 648.794062] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7596b5ff-7280-41a1-9a1c-9744d0ef91e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.814731] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 648.814731] env[65522]: value = "task-5113581" [ 648.814731] env[65522]: _type = "Task" [ 648.814731] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.828432] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113581, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.852599] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 648.852843] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 648.854796] env[65522]: INFO nova.compute.claims [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.967597] env[65522]: DEBUG oslo_vmware.api [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Task: {'id': task-5113577, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.564833} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.967904] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 648.968100] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 648.968314] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 648.968438] env[65522]: INFO nova.compute.manager [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Took 1.74 seconds to destroy the instance on the hypervisor. [ 648.968679] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 648.968879] env[65522]: DEBUG nova.compute.manager [-] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 648.968975] env[65522]: DEBUG nova.network.neutron [-] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 648.969229] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.969747] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.970016] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.009704] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e6e3d3-822b-518f-8296-60c8e74d1b72, 'name': SearchDatastore_Task, 'duration_secs': 0.020481} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.010259] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 649.010531] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20/2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 649.010850] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6852f01e-e105-4c97-a9b1-707676a4c8ca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.020237] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Waiting for the task: (returnval){ [ 649.020237] env[65522]: value = "task-5113582" [ 649.020237] env[65522]: _type = "Task" [ 649.020237] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.029928] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113582, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.032435] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ba4d723-da21-45a5-94a2-f5cab6e26f48 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "7c6f8218-602d-44f3-8012-de5a96972785" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.318s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 649.051034] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113578, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.119732} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.051331] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 649.052152] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4cdfaa-5d21-48a3-97b3-f149fda19483 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.056103] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 649.056363] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e112148d-7624-43ef-bf80-b7dfcf855c2a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.076121] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Reconfiguring VM instance instance-0000000b to attach disk [datastore2] 3f749f68-879f-44f3-8477-7ecea947078c/3f749f68-879f-44f3-8477-7ecea947078c.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 649.076898] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f14b3d7-b7e3-43f7-89f0-f20fa9ec312e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.094339] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 649.094339] env[65522]: value = "task-5113583" [ 649.094339] env[65522]: _type = "Task" [ 649.094339] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.100882] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Waiting for the task: (returnval){ [ 649.100882] env[65522]: value = "task-5113584" [ 649.100882] env[65522]: _type = "Task" [ 649.100882] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.104544] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113583, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.114428] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113584, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.260138] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.327112] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113581, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.511136] env[65522]: DEBUG nova.network.neutron [-] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 649.536787] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113582, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.615620] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113583, 'name': PowerOffVM_Task, 'duration_secs': 0.322258} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.616700] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 649.617013] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 649.618035] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c9cd72-d98f-41db-9d6c-eb33f8191b4c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.627698] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113584, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.638299] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 649.638299] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f958c84-ff12-44e5-8f03-57566a010600 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.641872] env[65522]: WARNING openstack [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.642336] env[65522]: WARNING openstack [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.736309] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 649.736648] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 649.737221] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleting the datastore file [datastore1] c93b52c3-1143-44d7-8c10-ac58bc25f541 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 649.737221] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-943a34f7-d122-47ae-86f8-6151b894634b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.749256] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 649.749256] env[65522]: value = "task-5113586" [ 649.749256] env[65522]: _type = "Task" [ 649.749256] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.758044] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113586, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.769935] env[65522]: WARNING neutronclient.v2_0.client [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.770530] env[65522]: WARNING openstack [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.770835] env[65522]: WARNING openstack [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.829025] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113581, 'name': CreateVM_Task, 'duration_secs': 0.653274} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.829025] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 649.829025] env[65522]: WARNING neutronclient.v2_0.client [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.829025] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.829025] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 649.829324] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 649.829324] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8bd1d35-bd35-4295-bac6-54462eccc513 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.835536] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 649.835536] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52803fc9-0164-b43c-a6b4-7489124237c7" [ 649.835536] env[65522]: _type = "Task" [ 649.835536] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.852918] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52803fc9-0164-b43c-a6b4-7489124237c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.900285] env[65522]: DEBUG nova.network.neutron [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updated VIF entry in instance network info cache for port cf16c93a-755c-435b-8c86-833d8ea8afbf. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 649.900285] env[65522]: DEBUG nova.network.neutron [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance_info_cache with network_info: [{"id": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "address": "fa:16:3e:69:05:d7", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf16c93a-75", "ovs_interfaceid": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 650.019023] env[65522]: INFO nova.compute.manager [-] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Took 1.54 seconds to deallocate network for instance. [ 650.030245] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 650.030572] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 650.047428] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113582, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.680717} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.050780] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20/2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 650.050780] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 650.050780] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-676ff53c-f9b6-48aa-841c-d16c2e533fef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.058778] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Waiting for the task: (returnval){ [ 650.058778] env[65522]: value = "task-5113587" [ 650.058778] env[65522]: _type = "Task" [ 650.058778] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.075197] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113587, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.124025] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113584, 'name': ReconfigVM_Task, 'duration_secs': 0.627775} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.125333] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Reconfigured VM instance instance-0000000b to attach disk [datastore2] 3f749f68-879f-44f3-8477-7ecea947078c/3f749f68-879f-44f3-8477-7ecea947078c.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 650.125333] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a9248af-b8e7-4f83-8af6-e4ef00ac5826 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.135972] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Waiting for the task: (returnval){ [ 650.135972] env[65522]: value = "task-5113588" [ 650.135972] env[65522]: _type = "Task" [ 650.135972] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.144831] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113588, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.186374] env[65522]: DEBUG nova.network.neutron [-] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 650.191514] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d3b634-b884-4678-b84e-f3150bc4e751 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.202055] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac486f51-c20f-47de-bc97-04574025c264 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.245650] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42635d73-db16-4e7c-a057-d369fdf0673c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.263631] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81af2269-a9fd-43e9-9e34-9adb840f4beb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.267723] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113586, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.355288} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.268112] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 650.268292] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 650.268501] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 650.283314] env[65522]: DEBUG nova.compute.provider_tree [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.346699] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52803fc9-0164-b43c-a6b4-7489124237c7, 'name': SearchDatastore_Task, 'duration_secs': 0.014778} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.347746] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 650.347746] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 650.347746] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.347923] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 650.348193] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 650.348534] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de15dd4d-8d85-4867-b615-0e589e14a82e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.380042] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 650.380317] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 650.381143] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1b71ecb-086e-4542-8efe-c4c9ea87a6b7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.388794] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 650.388794] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5257f871-2db1-5a6e-5535-cdcb74c09578" [ 650.388794] env[65522]: _type = "Task" [ 650.388794] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.398434] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5257f871-2db1-5a6e-5535-cdcb74c09578, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.403198] env[65522]: DEBUG oslo_concurrency.lockutils [req-ba593cd8-5f99-4b02-8330-61f08ad797a1 req-a104f0c1-295b-47bb-9b32-5825222770a4 service nova] Releasing lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 650.534670] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 650.542833] env[65522]: DEBUG nova.compute.manager [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 650.572788] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113587, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.650103] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113588, 'name': Rename_Task, 'duration_secs': 0.28283} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.650809] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 650.653019] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5f02aab-b0f9-4ac5-82d0-60398426ca8a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.662027] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Waiting for the task: (returnval){ [ 650.662027] env[65522]: value = "task-5113589" [ 650.662027] env[65522]: _type = "Task" [ 650.662027] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.674411] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113589, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.689604] env[65522]: INFO nova.compute.manager [-] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Took 1.72 seconds to deallocate network for instance. [ 650.787188] env[65522]: DEBUG nova.scheduler.client.report [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 650.882922] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Acquiring lock "fdf56fa6-ad18-499e-a171-6242abb293bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 650.883587] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Lock "fdf56fa6-ad18-499e-a171-6242abb293bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 650.903836] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5257f871-2db1-5a6e-5535-cdcb74c09578, 'name': SearchDatastore_Task, 'duration_secs': 0.0732} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.905891] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79a237dc-02ee-4cec-bb5e-6981818593fa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.913649] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 650.913649] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f6672b-44e5-44cd-41d4-b6921099b91f" [ 650.913649] env[65522]: _type = "Task" [ 650.913649] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.929096] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f6672b-44e5-44cd-41d4-b6921099b91f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.069254] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 651.074375] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113587, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.818556} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.074673] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 651.075455] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b48e12a1-e7ba-4f10-a62e-ba531c7bee75 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.105706] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Reconfiguring VM instance instance-0000000c to attach disk [datastore1] 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20/2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 651.106943] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4af479bc-97e6-4791-9a73-ddc3f80036f5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.130009] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Waiting for the task: (returnval){ [ 651.130009] env[65522]: value = "task-5113590" [ 651.130009] env[65522]: _type = "Task" [ 651.130009] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.144553] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113590, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.173215] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113589, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.200840] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 651.300328] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 651.300953] env[65522]: DEBUG nova.compute.manager [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 651.303596] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.769s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 651.303887] env[65522]: DEBUG nova.objects.instance [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Lazy-loading 'resources' on Instance uuid dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 651.311069] env[65522]: DEBUG nova.virt.hardware [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 651.311728] env[65522]: DEBUG nova.virt.hardware [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 651.311728] env[65522]: DEBUG nova.virt.hardware [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 651.311908] env[65522]: DEBUG nova.virt.hardware [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 651.311908] env[65522]: DEBUG nova.virt.hardware [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 651.312205] env[65522]: DEBUG nova.virt.hardware [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 651.312385] env[65522]: DEBUG nova.virt.hardware [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.312584] env[65522]: DEBUG nova.virt.hardware [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 651.312825] env[65522]: DEBUG nova.virt.hardware [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 651.313058] env[65522]: DEBUG nova.virt.hardware [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 651.313279] env[65522]: DEBUG nova.virt.hardware [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 651.314284] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54fd3853-4d08-4b06-827d-9d924c520bbb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.324751] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7f38f4-00a6-4c4c-ae79-73b233094cf3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.341490] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:04:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6fb0104-186b-4288-b87e-634893f46f01', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '573a1035-e00a-4cd2-a067-934487b65b60', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 651.349765] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 651.349765] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 651.350103] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-99991abd-900c-42c4-aa39-e423a3720c7d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.372263] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 651.372263] env[65522]: value = "task-5113591" [ 651.372263] env[65522]: _type = "Task" [ 651.372263] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.384118] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113591, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.386890] env[65522]: DEBUG nova.compute.manager [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 651.425768] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f6672b-44e5-44cd-41d4-b6921099b91f, 'name': SearchDatastore_Task, 'duration_secs': 0.027463} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.426514] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 651.426661] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 6c8165f5-6769-4954-b9ac-de8551954377/6c8165f5-6769-4954-b9ac-de8551954377.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 651.427073] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d5e3ba46-e9f2-47e4-89c3-86b7368aa132 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.438978] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 651.438978] env[65522]: value = "task-5113592" [ 651.438978] env[65522]: _type = "Task" [ 651.438978] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.450801] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113592, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.644454] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.676128] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113589, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.807639] env[65522]: DEBUG nova.compute.utils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 651.813526] env[65522]: DEBUG nova.compute.manager [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 651.813526] env[65522]: DEBUG nova.network.neutron [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 651.813836] env[65522]: WARNING neutronclient.v2_0.client [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 651.814319] env[65522]: WARNING neutronclient.v2_0.client [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 651.814908] env[65522]: WARNING openstack [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.815372] env[65522]: WARNING openstack [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.824170] env[65522]: DEBUG nova.compute.manager [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 651.897162] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113591, 'name': CreateVM_Task, 'duration_secs': 0.390662} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.903897] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 651.905306] env[65522]: WARNING neutronclient.v2_0.client [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 651.905794] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.906097] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 651.906476] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 651.907154] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-368dc315-f640-473a-b7f5-1bffc67906d4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.913916] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 651.913916] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523213b8-03e8-b68b-2a3a-3d5d6189d3b5" [ 651.913916] env[65522]: _type = "Task" [ 651.913916] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.921771] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 651.929129] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523213b8-03e8-b68b-2a3a-3d5d6189d3b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.949367] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113592, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.090123] env[65522]: DEBUG nova.compute.manager [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 652.090621] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7bbdd9e-4dbc-4915-b95a-3323f8b1bb3f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.096058] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be144f6b-9ba1-4a87-b98a-59f35de3fb98 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.107084] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-446af8d4-d8c9-4581-a6a2-cb0e9a7af66a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.164750] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed0357e-5bf8-4ee4-a7eb-55fa7f796cef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.186952] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abafa897-3110-4455-a671-ba3b744cbd4d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.191063] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113589, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.207243] env[65522]: DEBUG nova.compute.provider_tree [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.397897] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.427392] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523213b8-03e8-b68b-2a3a-3d5d6189d3b5, 'name': SearchDatastore_Task, 'duration_secs': 0.013664} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.428079] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 652.428461] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 652.428808] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.429062] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 652.429360] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 652.430348] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da6a4f34-5c8a-4fe4-b4cc-68416d27e46e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.443029] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 652.443029] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 652.445518] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43880038-39b8-4a54-aeb9-4b12611a5fb7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.452280] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 652.452280] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5210517b-42fc-7121-ce4f-29234740805b" [ 652.452280] env[65522]: _type = "Task" [ 652.452280] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.455992] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113592, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.010941} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.460072] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 6c8165f5-6769-4954-b9ac-de8551954377/6c8165f5-6769-4954-b9ac-de8551954377.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 652.460471] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 652.460958] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4b7c6ec5-4ebf-4dab-b849-27640f6a698f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.469765] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5210517b-42fc-7121-ce4f-29234740805b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.471211] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 652.471211] env[65522]: value = "task-5113593" [ 652.471211] env[65522]: _type = "Task" [ 652.471211] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.480712] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113593, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.489213] env[65522]: DEBUG nova.policy [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '226b224888c04908a8ac4e712efb5450', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4fdbd0ed78c41a591ccc0a587313ec8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 652.620195] env[65522]: INFO nova.compute.manager [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] instance snapshotting [ 652.623287] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54514fad-9848-4992-b1da-24cd1d56092c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.644493] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe57f81-d635-4185-8cda-dd06a6bb4b34 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.669939] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113590, 'name': ReconfigVM_Task, 'duration_secs': 1.43844} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.677492] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Reconfigured VM instance instance-0000000c to attach disk [datastore1] 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20/2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 652.681551] env[65522]: DEBUG nova.compute.manager [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Received event network-vif-plugged-366aac9d-7922-4057-b141-5f4d54f4e2ab {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 652.682309] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Acquiring lock "3f749f68-879f-44f3-8477-7ecea947078c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 652.682309] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Lock "3f749f68-879f-44f3-8477-7ecea947078c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 652.682309] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Lock "3f749f68-879f-44f3-8477-7ecea947078c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 652.682309] env[65522]: DEBUG nova.compute.manager [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] No waiting events found dispatching network-vif-plugged-366aac9d-7922-4057-b141-5f4d54f4e2ab {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 652.682722] env[65522]: WARNING nova.compute.manager [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Received unexpected event network-vif-plugged-366aac9d-7922-4057-b141-5f4d54f4e2ab for instance with vm_state building and task_state spawning. [ 652.682722] env[65522]: DEBUG nova.compute.manager [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Received event network-changed-366aac9d-7922-4057-b141-5f4d54f4e2ab {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 652.683605] env[65522]: DEBUG nova.compute.manager [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Refreshing instance network info cache due to event network-changed-366aac9d-7922-4057-b141-5f4d54f4e2ab. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 652.683605] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Acquiring lock "refresh_cache-3f749f68-879f-44f3-8477-7ecea947078c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.683605] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Acquired lock "refresh_cache-3f749f68-879f-44f3-8477-7ecea947078c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 652.683766] env[65522]: DEBUG nova.network.neutron [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Refreshing network info cache for port 366aac9d-7922-4057-b141-5f4d54f4e2ab {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 652.685769] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e62c9e6-171e-4991-8a1d-8466085b1bd8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.694628] env[65522]: DEBUG oslo_vmware.api [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113589, 'name': PowerOnVM_Task, 'duration_secs': 1.533654} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.696394] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 652.696394] env[65522]: INFO nova.compute.manager [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Took 12.54 seconds to spawn the instance on the hypervisor. [ 652.696394] env[65522]: DEBUG nova.compute.manager [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 652.698394] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Waiting for the task: (returnval){ [ 652.698394] env[65522]: value = "task-5113594" [ 652.698394] env[65522]: _type = "Task" [ 652.698394] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.698394] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d51247-d917-443a-8ebc-f861db334343 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.720220] env[65522]: DEBUG nova.scheduler.client.report [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 652.724275] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113594, 'name': Rename_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.839765] env[65522]: DEBUG nova.compute.manager [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 652.884233] env[65522]: DEBUG nova.virt.hardware [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 652.884761] env[65522]: DEBUG nova.virt.hardware [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 652.884761] env[65522]: DEBUG nova.virt.hardware [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 652.884894] env[65522]: DEBUG nova.virt.hardware [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 652.884958] env[65522]: DEBUG nova.virt.hardware [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 652.885157] env[65522]: DEBUG nova.virt.hardware [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 652.885334] env[65522]: DEBUG nova.virt.hardware [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 652.885764] env[65522]: DEBUG nova.virt.hardware [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 652.885764] env[65522]: DEBUG nova.virt.hardware [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 652.885861] env[65522]: DEBUG nova.virt.hardware [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 652.886499] env[65522]: DEBUG nova.virt.hardware [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 652.886957] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d44d758-d19d-4f08-8eea-7f618e453a23 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.901167] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f41ebd79-15ec-4c02-bb28-d74c52b922fb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.968537] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5210517b-42fc-7121-ce4f-29234740805b, 'name': SearchDatastore_Task, 'duration_secs': 0.028215} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.970168] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cade07b2-77d0-4675-a728-c30ab1750a73 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.985719] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 652.985719] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529d4692-0f27-a70c-202c-f7090a23174c" [ 652.985719] env[65522]: _type = "Task" [ 652.985719] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.986629] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113593, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.153536} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.986629] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 652.991719] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e3eae82-f1e7-4c68-bb3d-a09d20bd9e0c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.026247] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] 6c8165f5-6769-4954-b9ac-de8551954377/6c8165f5-6769-4954-b9ac-de8551954377.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 653.027078] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529d4692-0f27-a70c-202c-f7090a23174c, 'name': SearchDatastore_Task, 'duration_secs': 0.015536} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.028949] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc1a4103-53d5-4726-a1e5-923dbb113ebb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.048856] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 653.048856] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] c93b52c3-1143-44d7-8c10-ac58bc25f541/c93b52c3-1143-44d7-8c10-ac58bc25f541.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 653.052562] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6e857398-5470-4557-92e7-a8dcabe11d68 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.060934] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 653.060934] env[65522]: value = "task-5113596" [ 653.060934] env[65522]: _type = "Task" [ 653.060934] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.062470] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 653.062470] env[65522]: value = "task-5113595" [ 653.062470] env[65522]: _type = "Task" [ 653.062470] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.076274] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113596, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.079863] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113595, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.104235] env[65522]: DEBUG nova.network.neutron [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Successfully created port: aca7a84f-d08f-4c79-821e-1f90b00f141b {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 653.158756] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 653.158756] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-83f8f39e-a42f-48f3-afb2-44c16c12ea7c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.169165] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 653.169165] env[65522]: value = "task-5113597" [ 653.169165] env[65522]: _type = "Task" [ 653.169165] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.181975] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113597, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.186895] env[65522]: WARNING neutronclient.v2_0.client [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 653.188170] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 653.188530] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 653.212954] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113594, 'name': Rename_Task, 'duration_secs': 0.241803} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.213269] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 653.213521] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32e94e67-c68e-4651-b0fa-3b748bbb4c6b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.220611] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Waiting for the task: (returnval){ [ 653.220611] env[65522]: value = "task-5113598" [ 653.220611] env[65522]: _type = "Task" [ 653.220611] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.230441] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.926s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 653.240803] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.172s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 653.243792] env[65522]: INFO nova.compute.claims [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 653.247025] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113598, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.247560] env[65522]: INFO nova.compute.manager [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Took 19.24 seconds to build instance. [ 653.262909] env[65522]: INFO nova.scheduler.client.report [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Deleted allocations for instance dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7 [ 653.475597] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "0d52b207-ac69-48ce-b3af-7a83a499886e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 653.475856] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "0d52b207-ac69-48ce-b3af-7a83a499886e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 653.554830] env[65522]: DEBUG oslo_concurrency.lockutils [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Acquiring lock "bc40780e-43e2-4058-ab6f-d0d65e839128" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 653.556264] env[65522]: DEBUG oslo_concurrency.lockutils [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Lock "bc40780e-43e2-4058-ab6f-d0d65e839128" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 653.556548] env[65522]: DEBUG oslo_concurrency.lockutils [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Acquiring lock "bc40780e-43e2-4058-ab6f-d0d65e839128-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 653.556764] env[65522]: DEBUG oslo_concurrency.lockutils [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Lock "bc40780e-43e2-4058-ab6f-d0d65e839128-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 653.556913] env[65522]: DEBUG oslo_concurrency.lockutils [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Lock "bc40780e-43e2-4058-ab6f-d0d65e839128-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 653.560875] env[65522]: INFO nova.compute.manager [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Terminating instance [ 653.580098] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113596, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.587674] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113595, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.686260] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113597, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.740195] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113598, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.751816] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e65e508-6682-40c5-8005-883e79309607 tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Lock "3f749f68-879f-44f3-8477-7ecea947078c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.258s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 653.773951] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9e68302-fb8a-45d3-927a-2024c09f8b86 tempest-ServerExternalEventsTest-1570442889 tempest-ServerExternalEventsTest-1570442889-project-member] Lock "dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.010s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 653.982034] env[65522]: DEBUG nova.compute.manager [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 654.076022] env[65522]: DEBUG nova.compute.manager [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 654.076022] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 654.076249] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113596, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.080617] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e001f16-ad9c-482e-adb4-3adb8aa6d1dd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.089327] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 654.092686] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb55ad36-7852-4689-8463-190ef777ae05 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.094523] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113595, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.102744] env[65522]: DEBUG oslo_vmware.api [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Waiting for the task: (returnval){ [ 654.102744] env[65522]: value = "task-5113599" [ 654.102744] env[65522]: _type = "Task" [ 654.102744] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.115032] env[65522]: DEBUG oslo_vmware.api [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Task: {'id': task-5113599, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.185582] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113597, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.238950] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113598, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.505025] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 654.533396] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c24f54-efe7-43f3-9186-44abd3e367c7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.542844] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d378757f-017e-4c85-b5e3-79cfe94219b5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.591466] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fca5dd-10bd-4d36-bbbf-6da3c799e8a3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.601534] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113596, 'name': ReconfigVM_Task, 'duration_secs': 1.510024} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.607610] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Reconfigured VM instance instance-0000000d to attach disk [datastore2] 6c8165f5-6769-4954-b9ac-de8551954377/6c8165f5-6769-4954-b9ac-de8551954377.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 654.608662] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113595, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.398546} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.608662] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd9d135b-53b2-414f-8632-0c3810422037 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.614061] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77147652-843e-4a95-967c-9d67068dea42 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.619877] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] c93b52c3-1143-44d7-8c10-ac58bc25f541/c93b52c3-1143-44d7-8c10-ac58bc25f541.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 654.620142] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 654.620433] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-48a8da12-acf7-422a-b8b7-797014b226bd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.631451] env[65522]: DEBUG oslo_vmware.api [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Task: {'id': task-5113599, 'name': PowerOffVM_Task, 'duration_secs': 0.274375} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.643032] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 654.643324] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 654.643702] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 654.643702] env[65522]: value = "task-5113600" [ 654.643702] env[65522]: _type = "Task" [ 654.643702] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.644231] env[65522]: DEBUG nova.compute.provider_tree [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.645879] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 654.645879] env[65522]: value = "task-5113601" [ 654.645879] env[65522]: _type = "Task" [ 654.645879] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.646904] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56134e67-cb91-42f9-8bd7-ed4ad2e40807 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.661918] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113600, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.683990] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113597, 'name': CreateSnapshot_Task, 'duration_secs': 1.058577} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.683990] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 654.684869] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acac3f4e-ebb9-4f8e-96cb-6a160d3ecec1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.728959] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 654.733021] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 654.733021] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Deleting the datastore file [datastore2] bc40780e-43e2-4058-ab6f-d0d65e839128 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 654.733021] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5a792e1d-230f-4742-b835-2cab1b2e8ff5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.737321] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113598, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.746260] env[65522]: DEBUG oslo_vmware.api [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Waiting for the task: (returnval){ [ 654.746260] env[65522]: value = "task-5113603" [ 654.746260] env[65522]: _type = "Task" [ 654.746260] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.756547] env[65522]: DEBUG oslo_vmware.api [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Task: {'id': task-5113603, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.761528] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 654.761929] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 654.875721] env[65522]: WARNING neutronclient.v2_0.client [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 654.876412] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 654.877248] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.002263] env[65522]: DEBUG nova.network.neutron [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Successfully updated port: aca7a84f-d08f-4c79-821e-1f90b00f141b {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 655.044275] env[65522]: DEBUG nova.network.neutron [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Updated VIF entry in instance network info cache for port 366aac9d-7922-4057-b141-5f4d54f4e2ab. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 655.044645] env[65522]: DEBUG nova.network.neutron [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Updating instance_info_cache with network_info: [{"id": "366aac9d-7922-4057-b141-5f4d54f4e2ab", "address": "fa:16:3e:66:0d:26", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.217", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap366aac9d-79", "ovs_interfaceid": "366aac9d-7922-4057-b141-5f4d54f4e2ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 655.150708] env[65522]: DEBUG nova.scheduler.client.report [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 655.179763] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113600, 'name': Rename_Task, 'duration_secs': 0.220833} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.183480] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 655.183480] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113601, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.296222} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.184766] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-25ec4a76-ebac-4464-8e29-67b9706e8d06 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.187110] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 655.189471] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb5867f-7136-4b6c-91cf-e86a55259f7a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.214975] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 655.224868] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Reconfiguring VM instance instance-00000003 to attach disk [datastore1] c93b52c3-1143-44d7-8c10-ac58bc25f541/c93b52c3-1143-44d7-8c10-ac58bc25f541.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 655.228224] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-295eb608-e305-468b-a695-e24b60f3725b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.233650] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-988a4e99-5943-4878-9997-b4511a878566 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.249777] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 655.249777] env[65522]: value = "task-5113604" [ 655.249777] env[65522]: _type = "Task" [ 655.249777] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.267720] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 655.267720] env[65522]: value = "task-5113605" [ 655.267720] env[65522]: _type = "Task" [ 655.267720] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.268126] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 655.268126] env[65522]: value = "task-5113606" [ 655.268126] env[65522]: _type = "Task" [ 655.268126] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.291719] env[65522]: DEBUG oslo_vmware.api [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Task: {'id': task-5113603, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.219031} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.291719] env[65522]: DEBUG oslo_vmware.api [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5113598, 'name': PowerOnVM_Task, 'duration_secs': 1.63065} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.293306] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113604, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.294400] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 655.294976] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 655.294976] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 655.295759] env[65522]: INFO nova.compute.manager [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Took 1.22 seconds to destroy the instance on the hypervisor. [ 655.295854] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 655.296308] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 655.296407] env[65522]: INFO nova.compute.manager [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Took 13.29 seconds to spawn the instance on the hypervisor. [ 655.296605] env[65522]: DEBUG nova.compute.manager [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 655.311597] env[65522]: DEBUG nova.compute.manager [-] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 655.311597] env[65522]: DEBUG nova.network.neutron [-] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 655.311597] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.314343] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.314343] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.321784] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7041ac13-ff9b-4485-98c4-0e7273f6eff0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.324536] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.325043] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113605, 'name': CloneVM_Task} progress is 11%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.365833] env[65522]: DEBUG nova.compute.manager [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Received event network-vif-plugged-2eec2687-b0a0-442d-8306-27ab523bf504 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 655.366711] env[65522]: DEBUG oslo_concurrency.lockutils [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] Acquiring lock "6c8165f5-6769-4954-b9ac-de8551954377-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.366711] env[65522]: DEBUG oslo_concurrency.lockutils [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] Lock "6c8165f5-6769-4954-b9ac-de8551954377-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.367588] env[65522]: DEBUG oslo_concurrency.lockutils [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] Lock "6c8165f5-6769-4954-b9ac-de8551954377-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 655.367588] env[65522]: DEBUG nova.compute.manager [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] No waiting events found dispatching network-vif-plugged-2eec2687-b0a0-442d-8306-27ab523bf504 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 655.368263] env[65522]: WARNING nova.compute.manager [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Received unexpected event network-vif-plugged-2eec2687-b0a0-442d-8306-27ab523bf504 for instance with vm_state building and task_state spawning. [ 655.369200] env[65522]: DEBUG nova.compute.manager [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Received event network-changed-2eec2687-b0a0-442d-8306-27ab523bf504 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 655.369427] env[65522]: DEBUG nova.compute.manager [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Refreshing instance network info cache due to event network-changed-2eec2687-b0a0-442d-8306-27ab523bf504. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 655.369695] env[65522]: DEBUG oslo_concurrency.lockutils [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] Acquiring lock "refresh_cache-6c8165f5-6769-4954-b9ac-de8551954377" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.369804] env[65522]: DEBUG oslo_concurrency.lockutils [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] Acquired lock "refresh_cache-6c8165f5-6769-4954-b9ac-de8551954377" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.369959] env[65522]: DEBUG nova.network.neutron [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Refreshing network info cache for port 2eec2687-b0a0-442d-8306-27ab523bf504 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 655.408842] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.509464] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "refresh_cache-ad074365-1443-481d-990f-479c523219fa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.509759] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquired lock "refresh_cache-ad074365-1443-481d-990f-479c523219fa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.510030] env[65522]: DEBUG nova.network.neutron [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 655.547093] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Releasing lock "refresh_cache-3f749f68-879f-44f3-8477-7ecea947078c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 655.547377] env[65522]: DEBUG nova.compute.manager [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Received event network-vif-plugged-ef361904-5296-49e1-aeb2-182cd4adb91b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 655.547612] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Acquiring lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.547824] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.548041] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 655.548201] env[65522]: DEBUG nova.compute.manager [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] No waiting events found dispatching network-vif-plugged-ef361904-5296-49e1-aeb2-182cd4adb91b {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 655.548379] env[65522]: WARNING nova.compute.manager [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Received unexpected event network-vif-plugged-ef361904-5296-49e1-aeb2-182cd4adb91b for instance with vm_state building and task_state spawning. [ 655.548538] env[65522]: DEBUG nova.compute.manager [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Received event network-changed-ef361904-5296-49e1-aeb2-182cd4adb91b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 655.548688] env[65522]: DEBUG nova.compute.manager [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Refreshing instance network info cache due to event network-changed-ef361904-5296-49e1-aeb2-182cd4adb91b. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 655.548873] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Acquiring lock "refresh_cache-2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.551202] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Acquired lock "refresh_cache-2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.551202] env[65522]: DEBUG nova.network.neutron [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Refreshing network info cache for port ef361904-5296-49e1-aeb2-182cd4adb91b {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 655.675502] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 655.675678] env[65522]: DEBUG nova.compute.manager [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 655.680192] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.479s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.680624] env[65522]: DEBUG nova.objects.instance [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Lazy-loading 'resources' on Instance uuid fbb035b5-4562-44e3-9b95-8cf85299bb43 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 655.785556] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113604, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.795680] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113606, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.800321] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113605, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.846696] env[65522]: INFO nova.compute.manager [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Took 20.29 seconds to build instance. [ 655.872973] env[65522]: WARNING neutronclient.v2_0.client [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.873930] env[65522]: WARNING openstack [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.874423] env[65522]: WARNING openstack [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.890378] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "a8f9f223-023a-48fc-9186-a360a78e27a9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.890901] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "a8f9f223-023a-48fc-9186-a360a78e27a9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.972164] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "97dc8cac-8241-4912-a3ed-689439ef7ff8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.973533] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "97dc8cac-8241-4912-a3ed-689439ef7ff8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 656.015416] env[65522]: WARNING openstack [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.016293] env[65522]: WARNING openstack [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.058164] env[65522]: WARNING neutronclient.v2_0.client [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 656.058697] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.059071] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.184652] env[65522]: DEBUG nova.compute.utils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 656.186200] env[65522]: DEBUG nova.compute.manager [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 656.186598] env[65522]: DEBUG nova.network.neutron [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 656.188501] env[65522]: WARNING neutronclient.v2_0.client [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 656.188501] env[65522]: WARNING neutronclient.v2_0.client [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 656.188501] env[65522]: WARNING openstack [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.188501] env[65522]: WARNING openstack [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.269338] env[65522]: DEBUG oslo_vmware.api [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113604, 'name': PowerOnVM_Task, 'duration_secs': 0.899546} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.269714] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 656.269980] env[65522]: INFO nova.compute.manager [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Took 11.76 seconds to spawn the instance on the hypervisor. [ 656.270792] env[65522]: DEBUG nova.compute.manager [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 656.271403] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155cc113-ed2c-4fd9-8753-4d7714b94a17 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.297089] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113605, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.300791] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113606, 'name': ReconfigVM_Task, 'duration_secs': 0.702157} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.301960] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Reconfigured VM instance instance-00000003 to attach disk [datastore1] c93b52c3-1143-44d7-8c10-ac58bc25f541/c93b52c3-1143-44d7-8c10-ac58bc25f541.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 656.301960] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6041d2d0-39bf-4acb-b093-b9146a8323d5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.307398] env[65522]: DEBUG nova.network.neutron [-] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 656.317774] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 656.317774] env[65522]: value = "task-5113607" [ 656.317774] env[65522]: _type = "Task" [ 656.317774] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.334039] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113607, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.349965] env[65522]: DEBUG oslo_concurrency.lockutils [None req-24899773-aed1-4b2a-81f2-e1dbffda7983 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.777s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 656.380729] env[65522]: DEBUG nova.network.neutron [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 656.394045] env[65522]: DEBUG nova.compute.manager [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 656.402091] env[65522]: WARNING openstack [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.402477] env[65522]: WARNING openstack [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.478024] env[65522]: DEBUG nova.compute.manager [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 656.580519] env[65522]: DEBUG nova.policy [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24bc4117f7884e7d8a9333607b1276e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '95bca233019f4e2f8bac0dd800392a69', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 656.583858] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11d1fca-afd7-4917-81a1-83c2e4245da8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.593890] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82800b27-edde-4ef2-8496-872a9c028c24 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.635277] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a32589a-6254-4fcd-9d0a-670ba0ec9ec9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.646042] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c0e99e-180b-4e5d-8fca-eaccbbad21df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.661158] env[65522]: DEBUG nova.compute.provider_tree [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.701143] env[65522]: DEBUG nova.compute.manager [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 656.781412] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113605, 'name': CloneVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.807112] env[65522]: INFO nova.compute.manager [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Took 20.09 seconds to build instance. [ 656.810318] env[65522]: INFO nova.compute.manager [-] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Took 1.50 seconds to deallocate network for instance. [ 656.830054] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113607, 'name': Rename_Task, 'duration_secs': 0.334976} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.832351] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 656.832944] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff842421-b19d-44cc-9edd-499d612e3650 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.841453] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 656.841453] env[65522]: value = "task-5113608" [ 656.841453] env[65522]: _type = "Task" [ 656.841453] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.851231] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113608, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.878608] env[65522]: WARNING openstack [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.879149] env[65522]: WARNING openstack [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.928043] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 657.007268] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 657.171204] env[65522]: DEBUG nova.scheduler.client.report [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 657.287465] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113605, 'name': CloneVM_Task, 'duration_secs': 1.57439} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.287954] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Created linked-clone VM from snapshot [ 657.289152] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd822d65-f091-473b-8b13-5ea350a5d5da {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.299314] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Uploading image 7cf6a02c-41b2-43b6-acd1-9041875c5534 {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 657.310298] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9da8b9d9-d543-4d71-9d04-c8a5a463bc4b tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "6c8165f5-6769-4954-b9ac-de8551954377" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.254s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.319304] env[65522]: DEBUG oslo_concurrency.lockutils [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 657.328118] env[65522]: DEBUG oslo_vmware.rw_handles [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 657.328118] env[65522]: value = "vm-994702" [ 657.328118] env[65522]: _type = "VirtualMachine" [ 657.328118] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 657.328387] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3f98d499-c2d9-4703-a74c-1329accdfd15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.331675] env[65522]: DEBUG nova.network.neutron [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Successfully created port: 3f4547a2-4a63-4803-a79f-eda70e48dd48 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 657.339812] env[65522]: DEBUG oslo_vmware.rw_handles [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lease: (returnval){ [ 657.339812] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a9cc2f-c426-32c3-80d5-13dbfc3cc4fc" [ 657.339812] env[65522]: _type = "HttpNfcLease" [ 657.339812] env[65522]: } obtained for exporting VM: (result){ [ 657.339812] env[65522]: value = "vm-994702" [ 657.339812] env[65522]: _type = "VirtualMachine" [ 657.339812] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 657.340328] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the lease: (returnval){ [ 657.340328] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a9cc2f-c426-32c3-80d5-13dbfc3cc4fc" [ 657.340328] env[65522]: _type = "HttpNfcLease" [ 657.340328] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 657.351766] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 657.351766] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a9cc2f-c426-32c3-80d5-13dbfc3cc4fc" [ 657.351766] env[65522]: _type = "HttpNfcLease" [ 657.351766] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 657.355365] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113608, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.357773] env[65522]: WARNING neutronclient.v2_0.client [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 657.358614] env[65522]: WARNING openstack [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.358949] env[65522]: WARNING openstack [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.370558] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.370919] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.440316] env[65522]: WARNING neutronclient.v2_0.client [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 657.441113] env[65522]: WARNING openstack [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.441499] env[65522]: WARNING openstack [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.676917] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.996s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.679204] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.757s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 657.681559] env[65522]: INFO nova.compute.claims [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.712868] env[65522]: DEBUG nova.compute.manager [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 657.727320] env[65522]: INFO nova.scheduler.client.report [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Deleted allocations for instance fbb035b5-4562-44e3-9b95-8cf85299bb43 [ 657.767210] env[65522]: DEBUG nova.virt.hardware [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 657.767210] env[65522]: DEBUG nova.virt.hardware [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 657.767210] env[65522]: DEBUG nova.virt.hardware [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 657.767504] env[65522]: DEBUG nova.virt.hardware [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 657.767504] env[65522]: DEBUG nova.virt.hardware [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 657.767504] env[65522]: DEBUG nova.virt.hardware [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 657.767504] env[65522]: DEBUG nova.virt.hardware [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 657.767504] env[65522]: DEBUG nova.virt.hardware [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 657.767731] env[65522]: DEBUG nova.virt.hardware [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 657.767731] env[65522]: DEBUG nova.virt.hardware [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 657.767731] env[65522]: DEBUG nova.virt.hardware [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 657.768921] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d986877-fe33-41e3-ad54-dc4e5bc0f56d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.780714] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872a7805-ad92-41a9-93d9-02f7e2c95999 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.855084] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 657.855084] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a9cc2f-c426-32c3-80d5-13dbfc3cc4fc" [ 657.855084] env[65522]: _type = "HttpNfcLease" [ 657.855084] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 657.855369] env[65522]: DEBUG oslo_vmware.api [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113608, 'name': PowerOnVM_Task, 'duration_secs': 0.933319} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.856608] env[65522]: DEBUG oslo_vmware.rw_handles [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 657.856608] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a9cc2f-c426-32c3-80d5-13dbfc3cc4fc" [ 657.856608] env[65522]: _type = "HttpNfcLease" [ 657.856608] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 657.856969] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 657.857624] env[65522]: DEBUG nova.compute.manager [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 657.858498] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96378b42-a38e-4354-ba24-9544494b39ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.862431] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373ba0aa-017b-4e02-8d0f-133afa10ff42 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.878346] env[65522]: DEBUG oslo_vmware.rw_handles [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293bac4-bb88-2978-fd75-f9088345af59/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 657.878533] env[65522]: DEBUG oslo_vmware.rw_handles [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293bac4-bb88-2978-fd75-f9088345af59/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 657.979953] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e698c5d8-2400-4560-a5a7-bb2c9ec546f2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.022991] env[65522]: DEBUG nova.network.neutron [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Updated VIF entry in instance network info cache for port 2eec2687-b0a0-442d-8306-27ab523bf504. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 658.023549] env[65522]: DEBUG nova.network.neutron [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Updating instance_info_cache with network_info: [{"id": "2eec2687-b0a0-442d-8306-27ab523bf504", "address": "fa:16:3e:0c:f9:44", "network": {"id": "472285e2-f5f3-4616-92f7-4c0144d833d4", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-945513036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2bd38f300424db78568e0b8d9982f27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2eec2687-b0", "ovs_interfaceid": "2eec2687-b0a0-442d-8306-27ab523bf504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 658.045019] env[65522]: WARNING neutronclient.v2_0.client [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 658.046207] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.046207] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.141713] env[65522]: DEBUG nova.network.neutron [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Updating instance_info_cache with network_info: [{"id": "aca7a84f-d08f-4c79-821e-1f90b00f141b", "address": "fa:16:3e:99:c9:27", "network": {"id": "6a60138b-7f03-4817-ac9a-1facdc1cc0cc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1344754077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4fdbd0ed78c41a591ccc0a587313ec8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f85835c8-5d0c-4b2f-97c4-6c4006580f79", "external-id": "nsx-vlan-transportzone-245", "segmentation_id": 245, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaca7a84f-d0", "ovs_interfaceid": "aca7a84f-d08f-4c79-821e-1f90b00f141b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 658.238320] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc0a08ff-d5a7-415c-92bc-bb69108d39d6 tempest-ImagesNegativeTestJSON-2086281897 tempest-ImagesNegativeTestJSON-2086281897-project-member] Lock "fbb035b5-4562-44e3-9b95-8cf85299bb43" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.515s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 658.397363] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 658.526228] env[65522]: DEBUG oslo_concurrency.lockutils [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] Releasing lock "refresh_cache-6c8165f5-6769-4954-b9ac-de8551954377" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 658.526514] env[65522]: DEBUG nova.compute.manager [req-f7764729-d9f3-4085-a954-49aff8084a5d req-3e87855d-e89a-4f3b-bd19-685a0dd3fc23 service nova] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Received event network-vif-deleted-e6e22697-3ece-4047-b378-b47c9c5de778 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 658.646337] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Releasing lock "refresh_cache-ad074365-1443-481d-990f-479c523219fa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 658.647769] env[65522]: DEBUG nova.compute.manager [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Instance network_info: |[{"id": "aca7a84f-d08f-4c79-821e-1f90b00f141b", "address": "fa:16:3e:99:c9:27", "network": {"id": "6a60138b-7f03-4817-ac9a-1facdc1cc0cc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1344754077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4fdbd0ed78c41a591ccc0a587313ec8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f85835c8-5d0c-4b2f-97c4-6c4006580f79", "external-id": "nsx-vlan-transportzone-245", "segmentation_id": 245, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaca7a84f-d0", "ovs_interfaceid": "aca7a84f-d08f-4c79-821e-1f90b00f141b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 658.647974] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:c9:27', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f85835c8-5d0c-4b2f-97c4-6c4006580f79', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aca7a84f-d08f-4c79-821e-1f90b00f141b', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 658.660242] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 658.660610] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad074365-1443-481d-990f-479c523219fa] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 658.661090] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9baedcb8-356b-442e-bcc0-5109a0f1e271 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.687437] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 658.687437] env[65522]: value = "task-5113610" [ 658.687437] env[65522]: _type = "Task" [ 658.687437] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.698252] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113610, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.009931] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e585e9-3909-43a4-a109-79825d1a29cd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.021842] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d92c5d8-bc1b-4272-a85d-86b6da684fea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.063496] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67806624-5740-4966-acf4-fe30be223da5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.072767] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a50dcea-35b9-4e45-bf94-22c10b4ed163 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.092361] env[65522]: DEBUG nova.compute.provider_tree [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.132492] env[65522]: DEBUG nova.network.neutron [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Successfully updated port: 3f4547a2-4a63-4803-a79f-eda70e48dd48 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 659.156889] env[65522]: DEBUG nova.network.neutron [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Updated VIF entry in instance network info cache for port ef361904-5296-49e1-aeb2-182cd4adb91b. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 659.156889] env[65522]: DEBUG nova.network.neutron [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Updating instance_info_cache with network_info: [{"id": "ef361904-5296-49e1-aeb2-182cd4adb91b", "address": "fa:16:3e:58:20:55", "network": {"id": "d13db148-3532-452c-8c04-2a9a32a8c145", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1279941651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7340f053e9784f30be39030d99fe698f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f969bd9-e040-4b9b-85b2-7c61231584ad", "external-id": "nsx-vlan-transportzone-995", "segmentation_id": 995, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef361904-52", "ovs_interfaceid": "ef361904-5296-49e1-aeb2-182cd4adb91b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 659.203757] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113610, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.369864] env[65522]: DEBUG nova.compute.manager [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Received event network-vif-deleted-43c29c73-562e-48ba-886e-8e42d9c04074 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 659.370138] env[65522]: DEBUG nova.compute.manager [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Received event network-changed-a002c980-b974-432b-aa4e-e8613c617e8e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 659.370302] env[65522]: DEBUG nova.compute.manager [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Refreshing instance network info cache due to event network-changed-a002c980-b974-432b-aa4e-e8613c617e8e. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 659.370562] env[65522]: DEBUG oslo_concurrency.lockutils [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] Acquiring lock "refresh_cache-a09c12be-0aaa-4934-a1b3-5af29feba8a6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.370806] env[65522]: DEBUG oslo_concurrency.lockutils [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] Acquired lock "refresh_cache-a09c12be-0aaa-4934-a1b3-5af29feba8a6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 659.370994] env[65522]: DEBUG nova.network.neutron [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Refreshing network info cache for port a002c980-b974-432b-aa4e-e8613c617e8e {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 659.594534] env[65522]: DEBUG nova.scheduler.client.report [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 659.636892] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "refresh_cache-e5d2cc50-4ee6-4e64-9b52-888968a717ca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.637469] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquired lock "refresh_cache-e5d2cc50-4ee6-4e64-9b52-888968a717ca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 659.639259] env[65522]: DEBUG nova.network.neutron [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 659.662931] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Releasing lock "refresh_cache-2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 659.662931] env[65522]: DEBUG nova.compute.manager [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Received event network-changed-d24163e8-4f08-4908-9307-c4edb0d4ffb7 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 659.662931] env[65522]: DEBUG nova.compute.manager [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Refreshing instance network info cache due to event network-changed-d24163e8-4f08-4908-9307-c4edb0d4ffb7. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 659.662931] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Acquiring lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.662931] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Acquired lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 659.663112] env[65522]: DEBUG nova.network.neutron [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Refreshing network info cache for port d24163e8-4f08-4908-9307-c4edb0d4ffb7 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 659.706197] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113610, 'name': CreateVM_Task, 'duration_secs': 0.559382} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.706197] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad074365-1443-481d-990f-479c523219fa] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 659.706452] env[65522]: WARNING neutronclient.v2_0.client [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 659.707022] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.707109] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 659.707642] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 659.708454] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe83962a-7a68-4e41-9df5-6e90939196fb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.718955] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 659.718955] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e6775f-683c-dc05-d412-54e6640e1d34" [ 659.718955] env[65522]: _type = "Task" [ 659.718955] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.731694] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e6775f-683c-dc05-d412-54e6640e1d34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.874619] env[65522]: WARNING neutronclient.v2_0.client [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 659.876383] env[65522]: WARNING openstack [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.877287] env[65522]: WARNING openstack [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.103188] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.423s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 660.103295] env[65522]: DEBUG nova.compute.manager [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 660.106665] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.602s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 660.108288] env[65522]: INFO nova.compute.claims [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 660.144148] env[65522]: WARNING openstack [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.146131] env[65522]: WARNING openstack [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.166601] env[65522]: WARNING neutronclient.v2_0.client [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 660.167300] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.167673] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.177787] env[65522]: WARNING openstack [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.178224] env[65522]: WARNING openstack [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.230864] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e6775f-683c-dc05-d412-54e6640e1d34, 'name': SearchDatastore_Task, 'duration_secs': 0.017027} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.231470] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 660.231587] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 660.231764] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.231960] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 660.232217] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 660.232607] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f6ef560-7998-40f8-89d3-cff3ddf13774 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.244844] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 660.245136] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 660.245917] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-209fe800-9435-4341-a2a8-e81689a7b401 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.254225] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 660.254225] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52993f04-580d-5e89-88dc-f8e24c219a9f" [ 660.254225] env[65522]: _type = "Task" [ 660.254225] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.268000] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52993f04-580d-5e89-88dc-f8e24c219a9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.308517] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "ca964440-5375-4aff-8b45-96fbe829dd16" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 660.308816] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "ca964440-5375-4aff-8b45-96fbe829dd16" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 660.320537] env[65522]: DEBUG nova.network.neutron [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 660.455834] env[65522]: WARNING openstack [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.456501] env[65522]: WARNING openstack [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.555559] env[65522]: WARNING neutronclient.v2_0.client [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 660.556295] env[65522]: WARNING openstack [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.556663] env[65522]: WARNING openstack [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.614886] env[65522]: DEBUG nova.compute.utils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 660.624548] env[65522]: DEBUG nova.compute.manager [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 660.624548] env[65522]: DEBUG nova.network.neutron [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 660.624548] env[65522]: WARNING neutronclient.v2_0.client [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 660.624548] env[65522]: WARNING neutronclient.v2_0.client [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 660.625371] env[65522]: WARNING openstack [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.626829] env[65522]: WARNING openstack [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.641295] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.641722] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.732309] env[65522]: WARNING neutronclient.v2_0.client [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 660.732309] env[65522]: WARNING openstack [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.732309] env[65522]: WARNING openstack [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.768986] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52993f04-580d-5e89-88dc-f8e24c219a9f, 'name': SearchDatastore_Task, 'duration_secs': 0.01618} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.770651] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbb05b2c-390e-4e84-a093-3d4572bd619a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.778407] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 660.778407] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526855d2-4eb1-f1aa-e52a-c82791fd3c2c" [ 660.778407] env[65522]: _type = "Task" [ 660.778407] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.787348] env[65522]: DEBUG nova.policy [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '213c3620dddd4af58352733b54b80f9a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f64db4fc2f3412986ef6fd1393e99c8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 660.794343] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526855d2-4eb1-f1aa-e52a-c82791fd3c2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.812603] env[65522]: DEBUG nova.compute.manager [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 660.847538] env[65522]: DEBUG nova.network.neutron [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Updated VIF entry in instance network info cache for port a002c980-b974-432b-aa4e-e8613c617e8e. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 660.848446] env[65522]: DEBUG nova.network.neutron [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Updating instance_info_cache with network_info: [{"id": "a002c980-b974-432b-aa4e-e8613c617e8e", "address": "fa:16:3e:d4:83:9e", "network": {"id": "65a80e4b-308a-433d-8c2d-4f21879d8ac2", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1688898810-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b1927d35d9154557a33422045152daa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f847601f-7479-48eb-842f-41f94eea8537", "external-id": "nsx-vlan-transportzone-35", "segmentation_id": 35, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa002c980-b9", "ovs_interfaceid": "a002c980-b974-432b-aa4e-e8613c617e8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 660.875085] env[65522]: DEBUG nova.compute.manager [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Received event network-changed-dcf081b2-7a3e-4efd-84fa-bbf035dfd690 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 660.875601] env[65522]: DEBUG nova.compute.manager [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Refreshing instance network info cache due to event network-changed-dcf081b2-7a3e-4efd-84fa-bbf035dfd690. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 660.875601] env[65522]: DEBUG oslo_concurrency.lockutils [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Acquiring lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.876946] env[65522]: DEBUG oslo_concurrency.lockutils [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Acquired lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 660.876946] env[65522]: DEBUG nova.network.neutron [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Refreshing network info cache for port dcf081b2-7a3e-4efd-84fa-bbf035dfd690 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 660.954407] env[65522]: WARNING neutronclient.v2_0.client [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 660.955105] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.955528] env[65522]: WARNING openstack [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.990071] env[65522]: DEBUG nova.network.neutron [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Updating instance_info_cache with network_info: [{"id": "3f4547a2-4a63-4803-a79f-eda70e48dd48", "address": "fa:16:3e:c2:94:99", "network": {"id": "6314277f-6ed5-45e5-950d-2683782849aa", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-246048775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "95bca233019f4e2f8bac0dd800392a69", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f4547a2-4a", "ovs_interfaceid": "3f4547a2-4a63-4803-a79f-eda70e48dd48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 661.124221] env[65522]: DEBUG nova.compute.manager [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 661.181315] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "cd82f2fd-d657-4ece-8e90-9a47072f5546" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 661.182588] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "cd82f2fd-d657-4ece-8e90-9a47072f5546" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 661.291698] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526855d2-4eb1-f1aa-e52a-c82791fd3c2c, 'name': SearchDatastore_Task, 'duration_secs': 0.015996} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.292023] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.292322] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] ad074365-1443-481d-990f-479c523219fa/ad074365-1443-481d-990f-479c523219fa.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 661.292607] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad82b414-62b4-4c4b-b241-04792c666822 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.304048] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 661.304048] env[65522]: value = "task-5113611" [ 661.304048] env[65522]: _type = "Task" [ 661.304048] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.324097] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113611, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.340256] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 661.346145] env[65522]: DEBUG nova.network.neutron [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Updated VIF entry in instance network info cache for port d24163e8-4f08-4908-9307-c4edb0d4ffb7. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 661.346145] env[65522]: DEBUG nova.network.neutron [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Updating instance_info_cache with network_info: [{"id": "d24163e8-4f08-4908-9307-c4edb0d4ffb7", "address": "fa:16:3e:e2:90:31", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd24163e8-4f", "ovs_interfaceid": "d24163e8-4f08-4908-9307-c4edb0d4ffb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 661.352322] env[65522]: DEBUG oslo_concurrency.lockutils [req-3cc08ac6-8942-4fbe-a23d-4118762fc651 req-36ca61ef-cb2c-4264-b318-5db0f8a3ac4b service nova] Releasing lock "refresh_cache-a09c12be-0aaa-4934-a1b3-5af29feba8a6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.380105] env[65522]: WARNING neutronclient.v2_0.client [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 661.380617] env[65522]: WARNING openstack [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.381019] env[65522]: WARNING openstack [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 661.390371] env[65522]: DEBUG nova.network.neutron [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Successfully created port: a7a9fee0-67af-4446-b838-0cd8922465b9 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 661.493886] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Releasing lock "refresh_cache-e5d2cc50-4ee6-4e64-9b52-888968a717ca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.494312] env[65522]: DEBUG nova.compute.manager [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Instance network_info: |[{"id": "3f4547a2-4a63-4803-a79f-eda70e48dd48", "address": "fa:16:3e:c2:94:99", "network": {"id": "6314277f-6ed5-45e5-950d-2683782849aa", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-246048775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "95bca233019f4e2f8bac0dd800392a69", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f4547a2-4a", "ovs_interfaceid": "3f4547a2-4a63-4803-a79f-eda70e48dd48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 661.495032] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:94:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '136c3499-9ca0-4f85-903d-1f194aa66ed9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3f4547a2-4a63-4803-a79f-eda70e48dd48', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 661.505457] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Creating folder: Project (95bca233019f4e2f8bac0dd800392a69). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 661.507605] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4243321e-2afc-485b-a46d-dfd6ccadb99c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.528368] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Created folder: Project (95bca233019f4e2f8bac0dd800392a69) in parent group-v994660. [ 661.528605] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Creating folder: Instances. Parent ref: group-v994704. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 661.532617] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b2a1b899-65c6-4fcd-abec-8df9c93093ac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.546221] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Created folder: Instances in parent group-v994704. [ 661.546652] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 661.546989] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 661.547304] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e2aaeff8-6ff3-44f3-bdfb-613e321c0a08 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.578448] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 661.578448] env[65522]: value = "task-5113614" [ 661.578448] env[65522]: _type = "Task" [ 661.578448] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.594039] env[65522]: WARNING openstack [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.594039] env[65522]: WARNING openstack [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 661.604174] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113614, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.619430] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98d08a5-4c04-473f-992b-7e62d12f9be0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.629323] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5667d4-37bc-4fe4-9994-21909fedbff8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.674702] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9282e82d-ee93-487e-803f-03cf05fa7e37 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.684600] env[65522]: DEBUG nova.compute.manager [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 661.688912] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e78978-d1b8-490e-a520-9b3e247df7b5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.708741] env[65522]: DEBUG nova.compute.provider_tree [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.817145] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113611, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.851893] env[65522]: DEBUG oslo_concurrency.lockutils [req-1c6ebdbd-2f2a-41b2-ac00-92c1838d0196 req-ad06ce19-3f21-46a7-bb83-c31ee57e2e8a service nova] Releasing lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.865125] env[65522]: WARNING neutronclient.v2_0.client [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 661.866332] env[65522]: WARNING openstack [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.867175] env[65522]: WARNING openstack [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.092433] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113614, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.145024] env[65522]: DEBUG nova.compute.manager [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 662.173020] env[65522]: DEBUG nova.virt.hardware [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 662.173020] env[65522]: DEBUG nova.virt.hardware [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 662.173020] env[65522]: DEBUG nova.virt.hardware [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 662.173020] env[65522]: DEBUG nova.virt.hardware [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 662.173328] env[65522]: DEBUG nova.virt.hardware [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 662.173328] env[65522]: DEBUG nova.virt.hardware [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 662.173328] env[65522]: DEBUG nova.virt.hardware [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 662.173328] env[65522]: DEBUG nova.virt.hardware [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 662.173328] env[65522]: DEBUG nova.virt.hardware [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 662.173547] env[65522]: DEBUG nova.virt.hardware [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 662.173547] env[65522]: DEBUG nova.virt.hardware [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 662.173719] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52e5540-22b2-47e6-adb3-41545b468d8d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.184023] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56723f7-3b2e-4c8c-b5bf-3ed61696de2e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.214236] env[65522]: DEBUG nova.scheduler.client.report [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 662.226236] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 662.317951] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113611, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.749336} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.318369] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] ad074365-1443-481d-990f-479c523219fa/ad074365-1443-481d-990f-479c523219fa.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 662.318658] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 662.319065] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4554a1b1-d312-4091-bf9d-ec6a687fe3c4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.329040] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 662.329040] env[65522]: value = "task-5113615" [ 662.329040] env[65522]: _type = "Task" [ 662.329040] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.340264] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113615, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.592933] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113614, 'name': CreateVM_Task, 'duration_secs': 0.836956} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.593348] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 662.593911] env[65522]: WARNING neutronclient.v2_0.client [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.595957] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.595957] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 662.595957] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 662.595957] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e765be2b-456e-4925-98da-2e5664496ee9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.604685] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 662.604685] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52063732-2aaf-11b5-9187-06efdbf168fd" [ 662.604685] env[65522]: _type = "Task" [ 662.604685] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.615956] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52063732-2aaf-11b5-9187-06efdbf168fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.634998] env[65522]: DEBUG nova.network.neutron [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Updated VIF entry in instance network info cache for port dcf081b2-7a3e-4efd-84fa-bbf035dfd690. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 662.635659] env[65522]: DEBUG nova.network.neutron [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Updating instance_info_cache with network_info: [{"id": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "address": "fa:16:3e:9b:22:d2", "network": {"id": "b1e15367-b861-4a6c-8f94-921ed2b0a66f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1304672897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a47babe780ec4ae7b6c40f4531f09a38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "79c2e589-f55b-4843-8d99-2e565be16706", "external-id": "nsx-vlan-transportzone-858", "segmentation_id": 858, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcf081b2-7a", "ovs_interfaceid": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 662.723974] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.617s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 662.724477] env[65522]: DEBUG nova.compute.manager [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 662.731772] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.801s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 662.731841] env[65522]: INFO nova.compute.claims [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 662.842312] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113615, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109178} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.842588] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 662.843415] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517e8a9b-f81c-4fec-a9de-fa27324b538f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.869684] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Reconfiguring VM instance instance-0000000e to attach disk [datastore1] ad074365-1443-481d-990f-479c523219fa/ad074365-1443-481d-990f-479c523219fa.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 662.870411] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6c5bce8-7478-4c63-93f5-cd11e39b6fef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.897374] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 662.897374] env[65522]: value = "task-5113616" [ 662.897374] env[65522]: _type = "Task" [ 662.897374] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.908877] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113616, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.114491] env[65522]: DEBUG nova.compute.manager [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Received event network-changed-cf16c93a-755c-435b-8c86-833d8ea8afbf {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 663.114683] env[65522]: DEBUG nova.compute.manager [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Refreshing instance network info cache due to event network-changed-cf16c93a-755c-435b-8c86-833d8ea8afbf. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 663.115102] env[65522]: DEBUG oslo_concurrency.lockutils [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Acquiring lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.115216] env[65522]: DEBUG oslo_concurrency.lockutils [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Acquired lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.115403] env[65522]: DEBUG nova.network.neutron [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Refreshing network info cache for port cf16c93a-755c-435b-8c86-833d8ea8afbf {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 663.128838] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52063732-2aaf-11b5-9187-06efdbf168fd, 'name': SearchDatastore_Task, 'duration_secs': 0.018286} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.129389] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 663.129389] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 663.130808] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.130808] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.130808] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 663.130808] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-25402047-cbe0-42a7-95a2-1d7a6323e5f5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.140753] env[65522]: DEBUG oslo_concurrency.lockutils [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Releasing lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 663.141060] env[65522]: DEBUG nova.compute.manager [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Received event network-vif-plugged-aca7a84f-d08f-4c79-821e-1f90b00f141b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 663.141277] env[65522]: DEBUG oslo_concurrency.lockutils [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Acquiring lock "ad074365-1443-481d-990f-479c523219fa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 663.141753] env[65522]: DEBUG oslo_concurrency.lockutils [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Lock "ad074365-1443-481d-990f-479c523219fa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 663.141753] env[65522]: DEBUG oslo_concurrency.lockutils [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Lock "ad074365-1443-481d-990f-479c523219fa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 663.142102] env[65522]: DEBUG nova.compute.manager [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] [instance: ad074365-1443-481d-990f-479c523219fa] No waiting events found dispatching network-vif-plugged-aca7a84f-d08f-4c79-821e-1f90b00f141b {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 663.142102] env[65522]: WARNING nova.compute.manager [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Received unexpected event network-vif-plugged-aca7a84f-d08f-4c79-821e-1f90b00f141b for instance with vm_state building and task_state spawning. [ 663.142102] env[65522]: DEBUG nova.compute.manager [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Received event network-changed-aca7a84f-d08f-4c79-821e-1f90b00f141b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 663.142236] env[65522]: DEBUG nova.compute.manager [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Refreshing instance network info cache due to event network-changed-aca7a84f-d08f-4c79-821e-1f90b00f141b. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 663.142397] env[65522]: DEBUG oslo_concurrency.lockutils [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Acquiring lock "refresh_cache-ad074365-1443-481d-990f-479c523219fa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.142955] env[65522]: DEBUG oslo_concurrency.lockutils [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Acquired lock "refresh_cache-ad074365-1443-481d-990f-479c523219fa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.142955] env[65522]: DEBUG nova.network.neutron [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Refreshing network info cache for port aca7a84f-d08f-4c79-821e-1f90b00f141b {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 663.148552] env[65522]: DEBUG nova.network.neutron [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Successfully updated port: a7a9fee0-67af-4446-b838-0cd8922465b9 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 663.155709] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 663.155850] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 663.157220] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc1df49c-0dd7-45e3-ae0a-ad53ba6709f1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.166786] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 663.166786] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529c25c8-9d60-6232-1d20-8bf5e188b3b7" [ 663.166786] env[65522]: _type = "Task" [ 663.166786] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.182560] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529c25c8-9d60-6232-1d20-8bf5e188b3b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.242649] env[65522]: DEBUG nova.compute.utils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 663.247228] env[65522]: DEBUG nova.compute.manager [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 663.247457] env[65522]: DEBUG nova.network.neutron [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 663.247798] env[65522]: WARNING neutronclient.v2_0.client [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.248207] env[65522]: WARNING neutronclient.v2_0.client [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.249042] env[65522]: WARNING openstack [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 663.249246] env[65522]: WARNING openstack [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 663.416105] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113616, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.474297] env[65522]: DEBUG nova.policy [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '149b8f2f034844f484632b8859ffd12b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '234a226c969b4965bfa2c411bb4c8d90', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 663.623818] env[65522]: WARNING neutronclient.v2_0.client [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.624577] env[65522]: WARNING openstack [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 663.624860] env[65522]: WARNING openstack [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 663.657778] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Acquiring lock "refresh_cache-fdf56fa6-ad18-499e-a171-6242abb293bf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.658108] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Acquired lock "refresh_cache-fdf56fa6-ad18-499e-a171-6242abb293bf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.658240] env[65522]: DEBUG nova.network.neutron [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 663.659817] env[65522]: WARNING neutronclient.v2_0.client [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.661286] env[65522]: WARNING openstack [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 663.661286] env[65522]: WARNING openstack [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 663.692548] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529c25c8-9d60-6232-1d20-8bf5e188b3b7, 'name': SearchDatastore_Task, 'duration_secs': 0.017431} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.693455] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b094846-0cff-4095-9e74-86739f0208fe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.700582] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 663.700582] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524d0ed9-198b-f67f-ab9e-11ae7974f96e" [ 663.700582] env[65522]: _type = "Task" [ 663.700582] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.711395] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524d0ed9-198b-f67f-ab9e-11ae7974f96e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.748344] env[65522]: DEBUG nova.compute.manager [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 663.920218] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113616, 'name': ReconfigVM_Task, 'duration_secs': 1.023108} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.920966] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Reconfigured VM instance instance-0000000e to attach disk [datastore1] ad074365-1443-481d-990f-479c523219fa/ad074365-1443-481d-990f-479c523219fa.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 663.921288] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e87d2997-abed-4de9-acdf-864457129e2f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.929624] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 663.929624] env[65522]: value = "task-5113617" [ 663.929624] env[65522]: _type = "Task" [ 663.929624] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.944265] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113617, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.166227] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6347321-abf0-4302-ace2-80f215238bf8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.176555] env[65522]: WARNING openstack [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.176843] env[65522]: WARNING openstack [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.186772] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef3026e-4c5c-4394-95e0-7c61807a16dc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.225567] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdba2783-2180-4311-93b1-0b8435827499 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.235381] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524d0ed9-198b-f67f-ab9e-11ae7974f96e, 'name': SearchDatastore_Task, 'duration_secs': 0.01547} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.238586] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 664.238894] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] e5d2cc50-4ee6-4e64-9b52-888968a717ca/e5d2cc50-4ee6-4e64-9b52-888968a717ca.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 664.239274] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5846912-1d01-47b0-9637-ec7363819b89 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.242591] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a59552a-9c26-49fc-9a49-677f474b8b7c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.262893] env[65522]: DEBUG nova.compute.provider_tree [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.265873] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 664.265873] env[65522]: value = "task-5113618" [ 664.265873] env[65522]: _type = "Task" [ 664.265873] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.279712] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113618, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.342046] env[65522]: DEBUG nova.network.neutron [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Successfully created port: 97dc7703-dd4a-43dd-8e2d-02f8a31d1a69 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 664.441714] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113617, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.559243] env[65522]: DEBUG nova.network.neutron [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 664.644043] env[65522]: WARNING openstack [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.644741] env[65522]: WARNING openstack [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.764508] env[65522]: DEBUG nova.compute.manager [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 664.774016] env[65522]: WARNING openstack [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.774439] env[65522]: WARNING openstack [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.782141] env[65522]: DEBUG nova.scheduler.client.report [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 664.787333] env[65522]: WARNING openstack [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.787333] env[65522]: WARNING openstack [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.801967] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113618, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.810829] env[65522]: DEBUG nova.virt.hardware [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:58:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='229151605',id=21,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-466502433',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 664.810963] env[65522]: DEBUG nova.virt.hardware [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 664.811748] env[65522]: DEBUG nova.virt.hardware [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 664.811748] env[65522]: DEBUG nova.virt.hardware [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 664.811748] env[65522]: DEBUG nova.virt.hardware [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 664.811899] env[65522]: DEBUG nova.virt.hardware [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 664.812023] env[65522]: DEBUG nova.virt.hardware [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 664.812180] env[65522]: DEBUG nova.virt.hardware [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 664.812333] env[65522]: DEBUG nova.virt.hardware [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 664.812490] env[65522]: DEBUG nova.virt.hardware [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 664.812655] env[65522]: DEBUG nova.virt.hardware [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 664.813863] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83bb270-44e0-4e79-843e-0dc7f5362253 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.823666] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8fb027-5091-4e52-80da-eab564cee618 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.944563] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113617, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.948442] env[65522]: WARNING neutronclient.v2_0.client [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 664.949130] env[65522]: WARNING openstack [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.949613] env[65522]: WARNING openstack [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.081369] env[65522]: WARNING neutronclient.v2_0.client [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.082059] env[65522]: WARNING openstack [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.086584] env[65522]: WARNING openstack [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.101026] env[65522]: WARNING neutronclient.v2_0.client [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.101407] env[65522]: WARNING openstack [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.102597] env[65522]: WARNING openstack [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.281588] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113618, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.941064} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.281878] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] e5d2cc50-4ee6-4e64-9b52-888968a717ca/e5d2cc50-4ee6-4e64-9b52-888968a717ca.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 665.282116] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 665.282523] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1a4fe716-d7ea-425c-922f-1fe5dfe012c8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.291880] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 665.291880] env[65522]: value = "task-5113619" [ 665.291880] env[65522]: _type = "Task" [ 665.291880] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.296579] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.567s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 665.297238] env[65522]: DEBUG nova.compute.manager [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 665.300400] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.293s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 665.302084] env[65522]: INFO nova.compute.claims [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 665.310598] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113619, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.357422] env[65522]: DEBUG nova.network.neutron [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updated VIF entry in instance network info cache for port cf16c93a-755c-435b-8c86-833d8ea8afbf. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 665.357815] env[65522]: DEBUG nova.network.neutron [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance_info_cache with network_info: [{"id": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "address": "fa:16:3e:69:05:d7", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf16c93a-75", "ovs_interfaceid": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 665.416908] env[65522]: DEBUG nova.compute.manager [req-de17b52b-d7e6-4a65-bb4f-a042e62c6127 req-12ef6a36-3cef-4bb7-bfc3-59202a9180db service nova] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Received event network-vif-plugged-a7a9fee0-67af-4446-b838-0cd8922465b9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 665.417151] env[65522]: DEBUG oslo_concurrency.lockutils [req-de17b52b-d7e6-4a65-bb4f-a042e62c6127 req-12ef6a36-3cef-4bb7-bfc3-59202a9180db service nova] Acquiring lock "fdf56fa6-ad18-499e-a171-6242abb293bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 665.417392] env[65522]: DEBUG oslo_concurrency.lockutils [req-de17b52b-d7e6-4a65-bb4f-a042e62c6127 req-12ef6a36-3cef-4bb7-bfc3-59202a9180db service nova] Lock "fdf56fa6-ad18-499e-a171-6242abb293bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 665.417509] env[65522]: DEBUG oslo_concurrency.lockutils [req-de17b52b-d7e6-4a65-bb4f-a042e62c6127 req-12ef6a36-3cef-4bb7-bfc3-59202a9180db service nova] Lock "fdf56fa6-ad18-499e-a171-6242abb293bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 665.418310] env[65522]: DEBUG nova.compute.manager [req-de17b52b-d7e6-4a65-bb4f-a042e62c6127 req-12ef6a36-3cef-4bb7-bfc3-59202a9180db service nova] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] No waiting events found dispatching network-vif-plugged-a7a9fee0-67af-4446-b838-0cd8922465b9 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 665.418568] env[65522]: WARNING nova.compute.manager [req-de17b52b-d7e6-4a65-bb4f-a042e62c6127 req-12ef6a36-3cef-4bb7-bfc3-59202a9180db service nova] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Received unexpected event network-vif-plugged-a7a9fee0-67af-4446-b838-0cd8922465b9 for instance with vm_state building and task_state spawning. [ 665.419500] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "87a3c63d-794d-44ab-bad6-65c323d72ae7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 665.419717] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "87a3c63d-794d-44ab-bad6-65c323d72ae7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 665.447291] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113617, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.463825] env[65522]: DEBUG nova.network.neutron [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Updated VIF entry in instance network info cache for port aca7a84f-d08f-4c79-821e-1f90b00f141b. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 665.464207] env[65522]: DEBUG nova.network.neutron [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Updating instance_info_cache with network_info: [{"id": "aca7a84f-d08f-4c79-821e-1f90b00f141b", "address": "fa:16:3e:99:c9:27", "network": {"id": "6a60138b-7f03-4817-ac9a-1facdc1cc0cc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1344754077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4fdbd0ed78c41a591ccc0a587313ec8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f85835c8-5d0c-4b2f-97c4-6c4006580f79", "external-id": "nsx-vlan-transportzone-245", "segmentation_id": 245, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaca7a84f-d0", "ovs_interfaceid": "aca7a84f-d08f-4c79-821e-1f90b00f141b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 665.504656] env[65522]: DEBUG nova.network.neutron [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Updating instance_info_cache with network_info: [{"id": "a7a9fee0-67af-4446-b838-0cd8922465b9", "address": "fa:16:3e:33:16:af", "network": {"id": "29f88cc3-e21a-4a6e-87a2-8624a1be6482", "bridge": "br-int", "label": "tempest-ServersTestJSON-1088062070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "5f64db4fc2f3412986ef6fd1393e99c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7a9fee0-67", "ovs_interfaceid": "a7a9fee0-67af-4446-b838-0cd8922465b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 665.804932] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113619, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083602} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.805306] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 665.807147] env[65522]: DEBUG nova.compute.utils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 665.813616] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c35048-dada-41e4-8e77-90368ea94aa1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.832122] env[65522]: DEBUG nova.compute.manager [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 665.834843] env[65522]: DEBUG nova.compute.manager [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 665.835416] env[65522]: DEBUG nova.network.neutron [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 665.835416] env[65522]: WARNING neutronclient.v2_0.client [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.836889] env[65522]: WARNING neutronclient.v2_0.client [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.836889] env[65522]: WARNING openstack [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.836889] env[65522]: WARNING openstack [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.873358] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Reconfiguring VM instance instance-0000000f to attach disk [datastore1] e5d2cc50-4ee6-4e64-9b52-888968a717ca/e5d2cc50-4ee6-4e64-9b52-888968a717ca.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 665.876712] env[65522]: DEBUG oslo_concurrency.lockutils [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Releasing lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 665.877206] env[65522]: DEBUG nova.compute.manager [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Received event network-vif-plugged-3f4547a2-4a63-4803-a79f-eda70e48dd48 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 665.877287] env[65522]: DEBUG oslo_concurrency.lockutils [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Acquiring lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 665.877577] env[65522]: DEBUG oslo_concurrency.lockutils [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 665.877769] env[65522]: DEBUG oslo_concurrency.lockutils [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 665.878081] env[65522]: DEBUG nova.compute.manager [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] No waiting events found dispatching network-vif-plugged-3f4547a2-4a63-4803-a79f-eda70e48dd48 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 665.878329] env[65522]: WARNING nova.compute.manager [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Received unexpected event network-vif-plugged-3f4547a2-4a63-4803-a79f-eda70e48dd48 for instance with vm_state building and task_state spawning. [ 665.880053] env[65522]: DEBUG nova.compute.manager [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Received event network-changed-3f4547a2-4a63-4803-a79f-eda70e48dd48 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 665.880053] env[65522]: DEBUG nova.compute.manager [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Refreshing instance network info cache due to event network-changed-3f4547a2-4a63-4803-a79f-eda70e48dd48. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 665.880053] env[65522]: DEBUG oslo_concurrency.lockutils [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Acquiring lock "refresh_cache-e5d2cc50-4ee6-4e64-9b52-888968a717ca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.880053] env[65522]: DEBUG oslo_concurrency.lockutils [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Acquired lock "refresh_cache-e5d2cc50-4ee6-4e64-9b52-888968a717ca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 665.880053] env[65522]: DEBUG nova.network.neutron [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Refreshing network info cache for port 3f4547a2-4a63-4803-a79f-eda70e48dd48 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 665.880699] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05911c33-81da-43ac-8fa7-b33024ef1aae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.897681] env[65522]: WARNING neutronclient.v2_0.client [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.898871] env[65522]: WARNING openstack [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.898871] env[65522]: WARNING openstack [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.917153] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 665.917153] env[65522]: value = "task-5113620" [ 665.917153] env[65522]: _type = "Task" [ 665.917153] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.923035] env[65522]: DEBUG nova.compute.manager [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 665.933970] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113620, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.947249] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113617, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.973303] env[65522]: DEBUG oslo_concurrency.lockutils [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] Releasing lock "refresh_cache-ad074365-1443-481d-990f-479c523219fa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 665.973303] env[65522]: DEBUG nova.compute.manager [req-eca65bbf-f398-48da-a957-de05a7f13e84 req-3c444b2f-316d-4ff5-92e2-821f14cf4b0b service nova] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Received event network-vif-deleted-9902660c-976f-4cee-8338-74f077c8f7b3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 666.009078] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Releasing lock "refresh_cache-fdf56fa6-ad18-499e-a171-6242abb293bf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 666.009500] env[65522]: DEBUG nova.compute.manager [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Instance network_info: |[{"id": "a7a9fee0-67af-4446-b838-0cd8922465b9", "address": "fa:16:3e:33:16:af", "network": {"id": "29f88cc3-e21a-4a6e-87a2-8624a1be6482", "bridge": "br-int", "label": "tempest-ServersTestJSON-1088062070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "5f64db4fc2f3412986ef6fd1393e99c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7a9fee0-67", "ovs_interfaceid": "a7a9fee0-67af-4446-b838-0cd8922465b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 666.009960] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:16:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '40859343-2baa-45fd-88e3-ebf8aaed2b19', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7a9fee0-67af-4446-b838-0cd8922465b9', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 666.018771] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Creating folder: Project (5f64db4fc2f3412986ef6fd1393e99c8). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 666.021755] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0b0bff70-66d3-4b64-b3c3-954862bba409 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.034329] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Created folder: Project (5f64db4fc2f3412986ef6fd1393e99c8) in parent group-v994660. [ 666.035222] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Creating folder: Instances. Parent ref: group-v994707. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 666.035222] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0c6323a6-6524-4bd5-9fe3-efddc860cf43 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.045989] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Created folder: Instances in parent group-v994707. [ 666.046324] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 666.046632] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 666.046889] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b64cfc74-1622-4ae9-8b8f-4b3bc1b4a431 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.075708] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 666.075708] env[65522]: value = "task-5113623" [ 666.075708] env[65522]: _type = "Task" [ 666.075708] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.087246] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113623, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.222731] env[65522]: DEBUG nova.policy [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2c898f8d63564df5928a08c323ec1561', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '70696b0fca8842ba8b8eac9cd984504f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 666.234401] env[65522]: DEBUG nova.network.neutron [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Successfully updated port: 97dc7703-dd4a-43dd-8e2d-02f8a31d1a69 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 666.254128] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b97a3c-0ca2-49c4-af9c-2b58acdae794 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.266043] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5904cf4-f8e0-4e68-bea5-7c772095c6a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.311063] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-476b8c69-893b-42d7-85cf-8afa44ec0f50 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.320773] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a31a510-7103-4157-a19c-be0aba1e61c2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.337277] env[65522]: DEBUG nova.compute.provider_tree [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.430708] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113620, 'name': ReconfigVM_Task, 'duration_secs': 0.417229} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.430973] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Reconfigured VM instance instance-0000000f to attach disk [datastore1] e5d2cc50-4ee6-4e64-9b52-888968a717ca/e5d2cc50-4ee6-4e64-9b52-888968a717ca.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 666.432029] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b87bcacb-436f-497d-8441-21cf2b7b386c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.446952] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 666.446952] env[65522]: value = "task-5113624" [ 666.446952] env[65522]: _type = "Task" [ 666.446952] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.450639] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113617, 'name': Rename_Task, 'duration_secs': 2.094867} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.455143] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 666.455399] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c9f1f4d-9081-4ba4-b681-4380f2c0b915 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.467225] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113624, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.469565] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.469565] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 666.469565] env[65522]: value = "task-5113625" [ 666.469565] env[65522]: _type = "Task" [ 666.469565] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.480246] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113625, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.523087] env[65522]: WARNING openstack [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.523729] env[65522]: WARNING openstack [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.588911] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113623, 'name': CreateVM_Task, 'duration_secs': 0.436865} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.588911] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 666.588911] env[65522]: WARNING neutronclient.v2_0.client [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.589322] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.589587] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 666.590067] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 666.590517] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ecf46cb-f5cb-4d7d-bb7b-47a4c2c4f5c0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.597740] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Waiting for the task: (returnval){ [ 666.597740] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c34016-4784-cbe6-96c8-fb276d87b5ad" [ 666.597740] env[65522]: _type = "Task" [ 666.597740] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.609138] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c34016-4784-cbe6-96c8-fb276d87b5ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.656017] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "c93b52c3-1143-44d7-8c10-ac58bc25f541" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.656266] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "c93b52c3-1143-44d7-8c10-ac58bc25f541" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 666.656482] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "c93b52c3-1143-44d7-8c10-ac58bc25f541-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.656674] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "c93b52c3-1143-44d7-8c10-ac58bc25f541-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 666.656849] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "c93b52c3-1143-44d7-8c10-ac58bc25f541-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 666.659298] env[65522]: INFO nova.compute.manager [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Terminating instance [ 666.737537] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "refresh_cache-0d52b207-ac69-48ce-b3af-7a83a499886e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.737790] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquired lock "refresh_cache-0d52b207-ac69-48ce-b3af-7a83a499886e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 666.737945] env[65522]: DEBUG nova.network.neutron [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 666.755383] env[65522]: DEBUG nova.network.neutron [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Successfully created port: 8cb15276-e492-4096-b65b-f1107c89c902 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 666.807906] env[65522]: WARNING neutronclient.v2_0.client [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.808714] env[65522]: WARNING openstack [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.809120] env[65522]: WARNING openstack [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.840816] env[65522]: DEBUG nova.scheduler.client.report [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 666.850844] env[65522]: DEBUG nova.compute.manager [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 666.894335] env[65522]: DEBUG nova.virt.hardware [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 666.895067] env[65522]: DEBUG nova.virt.hardware [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 666.895067] env[65522]: DEBUG nova.virt.hardware [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 666.895409] env[65522]: DEBUG nova.virt.hardware [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 666.895572] env[65522]: DEBUG nova.virt.hardware [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 666.895765] env[65522]: DEBUG nova.virt.hardware [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 666.895995] env[65522]: DEBUG nova.virt.hardware [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 666.896259] env[65522]: DEBUG nova.virt.hardware [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 666.896480] env[65522]: DEBUG nova.virt.hardware [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 666.896675] env[65522]: DEBUG nova.virt.hardware [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 666.896864] env[65522]: DEBUG nova.virt.hardware [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 666.898247] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4324d9e-38bd-4219-9128-bd9c751be7c5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.915745] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720c2eb8-ead2-4ed1-8c15-cdac5e2cbedd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.923745] env[65522]: DEBUG nova.compute.manager [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Received event network-changed-ef361904-5296-49e1-aeb2-182cd4adb91b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 666.923857] env[65522]: DEBUG nova.compute.manager [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Refreshing instance network info cache due to event network-changed-ef361904-5296-49e1-aeb2-182cd4adb91b. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 666.924051] env[65522]: DEBUG oslo_concurrency.lockutils [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] Acquiring lock "refresh_cache-2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.924606] env[65522]: DEBUG oslo_concurrency.lockutils [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] Acquired lock "refresh_cache-2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 666.924705] env[65522]: DEBUG nova.network.neutron [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Refreshing network info cache for port ef361904-5296-49e1-aeb2-182cd4adb91b {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 666.962517] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113624, 'name': Rename_Task, 'duration_secs': 0.196421} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.962786] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 666.963052] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5387b4f-7440-4daa-9f1c-244d63b1a1ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.971214] env[65522]: DEBUG nova.network.neutron [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Updated VIF entry in instance network info cache for port 3f4547a2-4a63-4803-a79f-eda70e48dd48. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 666.971556] env[65522]: DEBUG nova.network.neutron [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Updating instance_info_cache with network_info: [{"id": "3f4547a2-4a63-4803-a79f-eda70e48dd48", "address": "fa:16:3e:c2:94:99", "network": {"id": "6314277f-6ed5-45e5-950d-2683782849aa", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-246048775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "95bca233019f4e2f8bac0dd800392a69", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f4547a2-4a", "ovs_interfaceid": "3f4547a2-4a63-4803-a79f-eda70e48dd48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 666.979943] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 666.979943] env[65522]: value = "task-5113626" [ 666.979943] env[65522]: _type = "Task" [ 666.979943] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.988292] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113625, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.995525] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113626, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.112568] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c34016-4784-cbe6-96c8-fb276d87b5ad, 'name': SearchDatastore_Task, 'duration_secs': 0.018197} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.112568] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 667.113430] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 667.114413] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.114701] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 667.116265] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 667.116695] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4f050506-290d-40e5-8705-11c1631a98aa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.126922] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 667.127130] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 667.127897] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77b7340d-9d26-4f28-8106-8dfef0521b15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.134669] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Waiting for the task: (returnval){ [ 667.134669] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525bb484-6597-757f-6fa1-be1bdbb2d20b" [ 667.134669] env[65522]: _type = "Task" [ 667.134669] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.144260] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525bb484-6597-757f-6fa1-be1bdbb2d20b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.164793] env[65522]: DEBUG nova.compute.manager [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 667.164793] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 667.165353] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beff690a-bcaf-446e-81d9-f206bcdd9eaf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.173748] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 667.174495] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d76af7f-b844-42f6-91f3-a46a9fa7cee0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.181762] env[65522]: DEBUG oslo_vmware.api [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 667.181762] env[65522]: value = "task-5113627" [ 667.181762] env[65522]: _type = "Task" [ 667.181762] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.191825] env[65522]: DEBUG oslo_vmware.api [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113627, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.242933] env[65522]: WARNING openstack [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.243553] env[65522]: WARNING openstack [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.314913] env[65522]: DEBUG nova.network.neutron [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 667.338955] env[65522]: WARNING openstack [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.339476] env[65522]: WARNING openstack [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.348696] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.047s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 667.348696] env[65522]: DEBUG nova.compute.manager [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 667.351041] env[65522]: DEBUG oslo_concurrency.lockutils [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.032s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.351041] env[65522]: DEBUG nova.objects.instance [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Lazy-loading 'resources' on Instance uuid bc40780e-43e2-4058-ab6f-d0d65e839128 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 667.429827] env[65522]: WARNING neutronclient.v2_0.client [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.429827] env[65522]: WARNING openstack [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.430174] env[65522]: WARNING openstack [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.482572] env[65522]: WARNING neutronclient.v2_0.client [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.483602] env[65522]: WARNING openstack [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.484200] env[65522]: WARNING openstack [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.494584] env[65522]: DEBUG oslo_concurrency.lockutils [req-5311adaa-ce22-48db-a8ab-1394f4bc00c7 req-7149a773-510b-4e13-a79d-7dbaefcabc76 service nova] Releasing lock "refresh_cache-e5d2cc50-4ee6-4e64-9b52-888968a717ca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 667.514563] env[65522]: DEBUG oslo_vmware.api [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113625, 'name': PowerOnVM_Task, 'duration_secs': 0.654736} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.516029] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113626, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.518337] env[65522]: DEBUG oslo_vmware.rw_handles [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293bac4-bb88-2978-fd75-f9088345af59/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 667.520023] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 667.520243] env[65522]: INFO nova.compute.manager [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Took 14.68 seconds to spawn the instance on the hypervisor. [ 667.521200] env[65522]: DEBUG nova.compute.manager [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 667.521395] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb4ea04-6e0d-4445-bad7-9b6b76f049e1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.527177] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a67a02-9e7c-49d3-b838-0e7492b5a9df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.532072] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Acquiring lock "3f749f68-879f-44f3-8477-7ecea947078c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.532364] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Lock "3f749f68-879f-44f3-8477-7ecea947078c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.532605] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Acquiring lock "3f749f68-879f-44f3-8477-7ecea947078c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.532825] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Lock "3f749f68-879f-44f3-8477-7ecea947078c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.533014] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Lock "3f749f68-879f-44f3-8477-7ecea947078c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 667.535982] env[65522]: INFO nova.compute.manager [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Terminating instance [ 667.542866] env[65522]: DEBUG oslo_vmware.rw_handles [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293bac4-bb88-2978-fd75-f9088345af59/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 667.543140] env[65522]: ERROR oslo_vmware.rw_handles [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293bac4-bb88-2978-fd75-f9088345af59/disk-0.vmdk due to incomplete transfer. [ 667.550023] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-27e88480-e73a-4740-8212-dbdec6b5b730 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.555501] env[65522]: DEBUG oslo_vmware.rw_handles [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293bac4-bb88-2978-fd75-f9088345af59/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 667.555736] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Uploaded image 7cf6a02c-41b2-43b6-acd1-9041875c5534 to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 667.557817] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 667.559073] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a0230e2b-29d9-4fc4-9004-b3948b4d72dd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.573835] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 667.573835] env[65522]: value = "task-5113628" [ 667.573835] env[65522]: _type = "Task" [ 667.573835] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.586903] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113628, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.603626] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Acquiring lock "68d8d59b-4fc8-4993-8c9e-31b2329c88dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.603859] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Lock "68d8d59b-4fc8-4993-8c9e-31b2329c88dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.646284] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525bb484-6597-757f-6fa1-be1bdbb2d20b, 'name': SearchDatastore_Task, 'duration_secs': 0.01327} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.647123] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae796674-6115-40b8-8839-4bff6f2ffb01 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.655700] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Waiting for the task: (returnval){ [ 667.655700] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52934e32-c2fb-584f-9318-8f9a676e5570" [ 667.655700] env[65522]: _type = "Task" [ 667.655700] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.665234] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52934e32-c2fb-584f-9318-8f9a676e5570, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.693985] env[65522]: DEBUG oslo_vmware.api [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113627, 'name': PowerOffVM_Task, 'duration_secs': 0.230414} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.694492] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 667.694492] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 667.694705] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1bb9e658-d11d-4550-a522-d938bed2ca70 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.738352] env[65522]: DEBUG nova.network.neutron [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Updating instance_info_cache with network_info: [{"id": "97dc7703-dd4a-43dd-8e2d-02f8a31d1a69", "address": "fa:16:3e:16:68:0f", "network": {"id": "d0ade77e-db62-42ae-9058-c33fcab1d2f9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1207104058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "234a226c969b4965bfa2c411bb4c8d90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97dc7703-dd", "ovs_interfaceid": "97dc7703-dd4a-43dd-8e2d-02f8a31d1a69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 667.775271] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 667.775496] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 667.775672] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleting the datastore file [datastore1] c93b52c3-1143-44d7-8c10-ac58bc25f541 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 667.776066] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a99b4f46-a3e3-4e7f-9974-e95c84a9c89b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.785810] env[65522]: DEBUG oslo_vmware.api [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 667.785810] env[65522]: value = "task-5113630" [ 667.785810] env[65522]: _type = "Task" [ 667.785810] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.795210] env[65522]: DEBUG oslo_vmware.api [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113630, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.854514] env[65522]: DEBUG nova.compute.utils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 667.860698] env[65522]: DEBUG nova.compute.manager [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 667.860698] env[65522]: DEBUG nova.network.neutron [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 667.860698] env[65522]: WARNING neutronclient.v2_0.client [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.861212] env[65522]: WARNING neutronclient.v2_0.client [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.862833] env[65522]: WARNING openstack [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.863637] env[65522]: WARNING openstack [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.006752] env[65522]: DEBUG oslo_vmware.api [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113626, 'name': PowerOnVM_Task, 'duration_secs': 0.555431} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.009558] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 668.009765] env[65522]: INFO nova.compute.manager [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Took 10.30 seconds to spawn the instance on the hypervisor. [ 668.009957] env[65522]: DEBUG nova.compute.manager [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 668.011075] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e757a98e-a427-4f4b-9533-472dfa6e9db1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.025484] env[65522]: WARNING openstack [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.025861] env[65522]: WARNING openstack [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.047911] env[65522]: DEBUG nova.compute.manager [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 668.048277] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 668.054647] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e928c63-9efb-4d4d-946c-37078014d2ff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.064431] env[65522]: INFO nova.compute.manager [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Took 19.25 seconds to build instance. [ 668.071374] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 668.071540] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2a2b7f2-28d1-4eb3-85b3-db38d32e8477 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.083111] env[65522]: DEBUG oslo_vmware.api [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Waiting for the task: (returnval){ [ 668.083111] env[65522]: value = "task-5113631" [ 668.083111] env[65522]: _type = "Task" [ 668.083111] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.094654] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113628, 'name': Destroy_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.098361] env[65522]: DEBUG nova.policy [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24bc4117f7884e7d8a9333607b1276e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '95bca233019f4e2f8bac0dd800392a69', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 668.107181] env[65522]: DEBUG nova.compute.manager [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 668.110051] env[65522]: DEBUG oslo_vmware.api [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113631, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.169747] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52934e32-c2fb-584f-9318-8f9a676e5570, 'name': SearchDatastore_Task, 'duration_secs': 0.017582} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.169887] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 668.170186] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] fdf56fa6-ad18-499e-a171-6242abb293bf/fdf56fa6-ad18-499e-a171-6242abb293bf.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 668.170468] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d51ad0eb-3429-4bc6-8055-6dc5a4d1e293 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.179901] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Waiting for the task: (returnval){ [ 668.179901] env[65522]: value = "task-5113632" [ 668.179901] env[65522]: _type = "Task" [ 668.179901] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.191562] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113632, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.198277] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a801c5-73ce-4cbd-b3e6-3bc57315f128 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.208613] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a13af0f-9b25-4c8b-a518-60279dfaccbd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.244359] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Releasing lock "refresh_cache-0d52b207-ac69-48ce-b3af-7a83a499886e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 668.244775] env[65522]: DEBUG nova.compute.manager [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Instance network_info: |[{"id": "97dc7703-dd4a-43dd-8e2d-02f8a31d1a69", "address": "fa:16:3e:16:68:0f", "network": {"id": "d0ade77e-db62-42ae-9058-c33fcab1d2f9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1207104058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "234a226c969b4965bfa2c411bb4c8d90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97dc7703-dd", "ovs_interfaceid": "97dc7703-dd4a-43dd-8e2d-02f8a31d1a69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 668.250472] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:68:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '29945b49-d2b8-449d-9531-437917f49839', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '97dc7703-dd4a-43dd-8e2d-02f8a31d1a69', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 668.258525] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Creating folder: Project (234a226c969b4965bfa2c411bb4c8d90). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 668.259440] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30c93e7c-8898-46c1-855c-b8fe7ac1c43d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.263101] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-752663e0-4204-4c81-9574-adde4e300f53 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.273312] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1ca3e0-1636-432c-afe8-3b4cb8cbd9c3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.279652] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Created folder: Project (234a226c969b4965bfa2c411bb4c8d90) in parent group-v994660. [ 668.279862] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Creating folder: Instances. Parent ref: group-v994710. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 668.280995] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4a73c4bc-861a-4432-910b-fa52f5fb6faf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.294036] env[65522]: DEBUG nova.compute.provider_tree [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.307784] env[65522]: DEBUG oslo_vmware.api [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113630, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.448835} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.310618] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 668.310915] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 668.311378] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 668.311461] env[65522]: INFO nova.compute.manager [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Took 1.15 seconds to destroy the instance on the hypervisor. [ 668.311657] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 668.311874] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Created folder: Instances in parent group-v994710. [ 668.312084] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 668.312704] env[65522]: DEBUG nova.compute.manager [-] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 668.312790] env[65522]: DEBUG nova.network.neutron [-] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 668.312994] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.313625] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.313921] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.322376] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 668.322440] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2cf15d1f-3298-4397-9354-f04ef4d78ffe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.351998] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 668.351998] env[65522]: value = "task-5113635" [ 668.351998] env[65522]: _type = "Task" [ 668.351998] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.361987] env[65522]: DEBUG nova.compute.manager [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 668.364512] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113635, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.492963] env[65522]: DEBUG nova.network.neutron [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Successfully updated port: 8cb15276-e492-4096-b65b-f1107c89c902 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 668.527031] env[65522]: WARNING neutronclient.v2_0.client [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.528064] env[65522]: WARNING openstack [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.528611] env[65522]: WARNING openstack [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.554030] env[65522]: INFO nova.compute.manager [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Took 17.51 seconds to build instance. [ 668.567413] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5f224849-dc57-4606-b973-036818198ee2 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "ad074365-1443-481d-990f-479c523219fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.764s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 668.595641] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113628, 'name': Destroy_Task, 'duration_secs': 0.765857} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.596666] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Destroyed the VM [ 668.596780] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 668.597018] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9cfcd3d4-4bf2-48d7-a739-97a8eefdb0f4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.603289] env[65522]: DEBUG oslo_vmware.api [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113631, 'name': PowerOffVM_Task, 'duration_secs': 0.25862} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.604218] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 668.604218] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 668.606022] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e56552f7-f3b1-4c21-a0e1-1b41a6af38cd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.610380] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 668.610380] env[65522]: value = "task-5113636" [ 668.610380] env[65522]: _type = "Task" [ 668.610380] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.612488] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.633645] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113636, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.647039] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.670962] env[65522]: DEBUG nova.network.neutron [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Successfully created port: 475e049b-993f-4ab4-aba6-da16d2828c40 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 668.686302] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 668.686575] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 668.686752] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Deleting the datastore file [datastore2] 3f749f68-879f-44f3-8477-7ecea947078c {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 668.687709] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9dbc205f-5503-4204-86af-e2119fba41d4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.696682] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113632, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.698912] env[65522]: DEBUG oslo_vmware.api [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Waiting for the task: (returnval){ [ 668.698912] env[65522]: value = "task-5113638" [ 668.698912] env[65522]: _type = "Task" [ 668.698912] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.709941] env[65522]: DEBUG oslo_vmware.api [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113638, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.801654] env[65522]: DEBUG nova.scheduler.client.report [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 668.869257] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113635, 'name': CreateVM_Task, 'duration_secs': 0.472883} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.869525] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 668.871496] env[65522]: WARNING neutronclient.v2_0.client [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.871496] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.871496] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 668.871496] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 668.871496] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2050ff7a-2334-44af-8e99-8a665695f8d2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.877489] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 668.877489] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52557735-cefa-0165-90f3-4f54a896ebdf" [ 668.877489] env[65522]: _type = "Task" [ 668.877489] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.890963] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52557735-cefa-0165-90f3-4f54a896ebdf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.996833] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "refresh_cache-a8f9f223-023a-48fc-9186-a360a78e27a9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.996833] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquired lock "refresh_cache-a8f9f223-023a-48fc-9186-a360a78e27a9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 668.997089] env[65522]: DEBUG nova.network.neutron [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 669.057060] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2c3dca0-db6b-43e4-a7ec-04c23ef36d69 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.026s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.124848] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113636, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.174578] env[65522]: DEBUG nova.network.neutron [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Updated VIF entry in instance network info cache for port ef361904-5296-49e1-aeb2-182cd4adb91b. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 669.174951] env[65522]: DEBUG nova.network.neutron [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Updating instance_info_cache with network_info: [{"id": "ef361904-5296-49e1-aeb2-182cd4adb91b", "address": "fa:16:3e:58:20:55", "network": {"id": "d13db148-3532-452c-8c04-2a9a32a8c145", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1279941651-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7340f053e9784f30be39030d99fe698f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6f969bd9-e040-4b9b-85b2-7c61231584ad", "external-id": "nsx-vlan-transportzone-995", "segmentation_id": 995, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef361904-52", "ovs_interfaceid": "ef361904-5296-49e1-aeb2-182cd4adb91b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 669.189809] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113632, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.709988} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.190832] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] fdf56fa6-ad18-499e-a171-6242abb293bf/fdf56fa6-ad18-499e-a171-6242abb293bf.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 669.191280] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 669.191681] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-137b5655-6c9e-45c6-980a-cf5b1ca814fb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.200966] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Waiting for the task: (returnval){ [ 669.200966] env[65522]: value = "task-5113639" [ 669.200966] env[65522]: _type = "Task" [ 669.200966] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.214596] env[65522]: DEBUG oslo_vmware.api [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Task: {'id': task-5113638, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.342444} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.217779] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 669.217984] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 669.218206] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 669.218373] env[65522]: INFO nova.compute.manager [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Took 1.17 seconds to destroy the instance on the hypervisor. [ 669.218621] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 669.218822] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113639, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.219065] env[65522]: DEBUG nova.compute.manager [-] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 669.219200] env[65522]: DEBUG nova.network.neutron [-] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 669.219465] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 669.219989] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.220393] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.291619] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 669.311657] env[65522]: DEBUG oslo_concurrency.lockutils [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.961s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.314572] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 10.917s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.314839] env[65522]: DEBUG nova.objects.instance [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 669.342855] env[65522]: INFO nova.scheduler.client.report [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Deleted allocations for instance bc40780e-43e2-4058-ab6f-d0d65e839128 [ 669.373120] env[65522]: DEBUG nova.compute.manager [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 669.393027] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52557735-cefa-0165-90f3-4f54a896ebdf, 'name': SearchDatastore_Task, 'duration_secs': 0.026758} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.393027] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 669.393027] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 669.393027] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.393797] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 669.393797] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 669.393797] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7cc4020-680f-4313-a400-df986ba284c3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.407098] env[65522]: DEBUG nova.virt.hardware [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 669.407353] env[65522]: DEBUG nova.virt.hardware [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 669.407504] env[65522]: DEBUG nova.virt.hardware [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 669.407676] env[65522]: DEBUG nova.virt.hardware [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 669.407815] env[65522]: DEBUG nova.virt.hardware [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 669.408140] env[65522]: DEBUG nova.virt.hardware [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 669.408393] env[65522]: DEBUG nova.virt.hardware [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 669.408549] env[65522]: DEBUG nova.virt.hardware [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 669.408713] env[65522]: DEBUG nova.virt.hardware [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 669.408870] env[65522]: DEBUG nova.virt.hardware [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 669.409067] env[65522]: DEBUG nova.virt.hardware [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 669.409925] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db4f894-8ff4-45d8-bd45-da1dbcfc43b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.413627] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 669.413799] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 669.415062] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a32a501-f051-43cf-9f1b-f08677882917 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.421032] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2375287f-a336-4cae-8913-314f72d92d1b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.426243] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 669.426243] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c6a78a-5d05-8c62-c6f4-4f7f20926534" [ 669.426243] env[65522]: _type = "Task" [ 669.426243] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.444670] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c6a78a-5d05-8c62-c6f4-4f7f20926534, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.503537] env[65522]: WARNING openstack [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.503537] env[65522]: WARNING openstack [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.562661] env[65522]: DEBUG nova.network.neutron [-] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 669.627602] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113636, 'name': RemoveSnapshot_Task} progress is 81%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.677826] env[65522]: DEBUG oslo_concurrency.lockutils [req-5484f0da-eebc-4891-8341-d3b8bd2c8991 req-88710213-7fbd-4b11-9e97-5643f6cf8ee8 service nova] Releasing lock "refresh_cache-2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 669.719517] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113639, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079798} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.720127] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 669.721515] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ad1caa-30f2-4913-8a39-6435d1fa5cb0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.761231] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Reconfiguring VM instance instance-00000010 to attach disk [datastore1] fdf56fa6-ad18-499e-a171-6242abb293bf/fdf56fa6-ad18-499e-a171-6242abb293bf.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 669.761721] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d40efff-9056-4178-bb4e-c3df5816408e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.790248] env[65522]: DEBUG nova.network.neutron [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 669.799299] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Waiting for the task: (returnval){ [ 669.799299] env[65522]: value = "task-5113640" [ 669.799299] env[65522]: _type = "Task" [ 669.799299] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.815820] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113640, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.853328] env[65522]: DEBUG oslo_concurrency.lockutils [None req-184d2397-20ad-47c0-9c1e-339d28b2d336 tempest-DeleteServersAdminTestJSON-1776898441 tempest-DeleteServersAdminTestJSON-1776898441-project-admin] Lock "bc40780e-43e2-4058-ab6f-d0d65e839128" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.298s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.862946] env[65522]: WARNING openstack [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.863941] env[65522]: WARNING openstack [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.940560] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c6a78a-5d05-8c62-c6f4-4f7f20926534, 'name': SearchDatastore_Task, 'duration_secs': 0.028652} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.941529] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42721534-830f-4aa0-8900-f258bdd78315 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.946393] env[65522]: WARNING neutronclient.v2_0.client [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 669.947164] env[65522]: WARNING openstack [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.948057] env[65522]: WARNING openstack [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.957954] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 669.957954] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528b80ab-27c8-b2f9-33ba-78d106fcd75a" [ 669.957954] env[65522]: _type = "Task" [ 669.957954] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.967806] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528b80ab-27c8-b2f9-33ba-78d106fcd75a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.067665] env[65522]: INFO nova.compute.manager [-] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Took 1.75 seconds to deallocate network for instance. [ 670.128428] env[65522]: DEBUG oslo_vmware.api [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113636, 'name': RemoveSnapshot_Task, 'duration_secs': 1.392393} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.128712] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 670.128963] env[65522]: INFO nova.compute.manager [None req-efebeecb-4a24-481d-b5f1-1ea1b62256e1 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Took 17.51 seconds to snapshot the instance on the hypervisor. [ 670.151574] env[65522]: DEBUG nova.network.neutron [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Updating instance_info_cache with network_info: [{"id": "8cb15276-e492-4096-b65b-f1107c89c902", "address": "fa:16:3e:05:fb:2d", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cb15276-e4", "ovs_interfaceid": "8cb15276-e492-4096-b65b-f1107c89c902", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 670.153617] env[65522]: DEBUG nova.network.neutron [-] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 670.167899] env[65522]: DEBUG nova.compute.manager [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Received event network-changed-a7a9fee0-67af-4446-b838-0cd8922465b9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 670.168164] env[65522]: DEBUG nova.compute.manager [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Refreshing instance network info cache due to event network-changed-a7a9fee0-67af-4446-b838-0cd8922465b9. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 670.168360] env[65522]: DEBUG oslo_concurrency.lockutils [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] Acquiring lock "refresh_cache-fdf56fa6-ad18-499e-a171-6242abb293bf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.168893] env[65522]: DEBUG oslo_concurrency.lockutils [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] Acquired lock "refresh_cache-fdf56fa6-ad18-499e-a171-6242abb293bf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 670.169095] env[65522]: DEBUG nova.network.neutron [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Refreshing network info cache for port a7a9fee0-67af-4446-b838-0cd8922465b9 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 670.311063] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113640, 'name': ReconfigVM_Task, 'duration_secs': 0.494353} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.311458] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Reconfigured VM instance instance-00000010 to attach disk [datastore1] fdf56fa6-ad18-499e-a171-6242abb293bf/fdf56fa6-ad18-499e-a171-6242abb293bf.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 670.313021] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e56d2c4-d7a1-4768-8c61-95595be3d913 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.322213] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Waiting for the task: (returnval){ [ 670.322213] env[65522]: value = "task-5113641" [ 670.322213] env[65522]: _type = "Task" [ 670.322213] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.327751] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0a6a0529-94da-4504-b33c-dd58b8ac9236 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 670.329344] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.989s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 670.331734] env[65522]: INFO nova.compute.claims [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 670.342513] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113641, 'name': Rename_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.467263] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 670.467483] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 670.473611] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528b80ab-27c8-b2f9-33ba-78d106fcd75a, 'name': SearchDatastore_Task, 'duration_secs': 0.022428} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.474252] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 670.474521] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 0d52b207-ac69-48ce-b3af-7a83a499886e/0d52b207-ac69-48ce-b3af-7a83a499886e.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 670.474796] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-506f4275-2f82-4650-afaa-a21fae1a4076 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.482388] env[65522]: DEBUG nova.network.neutron [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Successfully updated port: 475e049b-993f-4ab4-aba6-da16d2828c40 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 670.486717] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 670.486717] env[65522]: value = "task-5113642" [ 670.486717] env[65522]: _type = "Task" [ 670.486717] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.497914] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113642, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.574247] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 670.656549] env[65522]: INFO nova.compute.manager [-] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Took 1.44 seconds to deallocate network for instance. [ 670.657131] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Releasing lock "refresh_cache-a8f9f223-023a-48fc-9186-a360a78e27a9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 670.658142] env[65522]: DEBUG nova.compute.manager [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Instance network_info: |[{"id": "8cb15276-e492-4096-b65b-f1107c89c902", "address": "fa:16:3e:05:fb:2d", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cb15276-e4", "ovs_interfaceid": "8cb15276-e492-4096-b65b-f1107c89c902", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 670.660675] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:fb:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8cb15276-e492-4096-b65b-f1107c89c902', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 670.667840] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Creating folder: Project (70696b0fca8842ba8b8eac9cd984504f). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 670.672198] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c7b45f14-c59e-4cff-93ce-18eb65d43b0f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.675094] env[65522]: WARNING neutronclient.v2_0.client [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 670.675743] env[65522]: WARNING openstack [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 670.676305] env[65522]: WARNING openstack [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 670.695170] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Created folder: Project (70696b0fca8842ba8b8eac9cd984504f) in parent group-v994660. [ 670.695420] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Creating folder: Instances. Parent ref: group-v994713. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 670.696180] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6b83a703-cb0c-430d-beae-2eec046a4273 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.707366] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Created folder: Instances in parent group-v994713. [ 670.707671] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 670.707881] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 670.709460] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8bae9141-2792-4cbb-8884-6f72481bd114 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.734824] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 670.734824] env[65522]: value = "task-5113645" [ 670.734824] env[65522]: _type = "Task" [ 670.734824] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.745523] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113645, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.836873] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113641, 'name': Rename_Task, 'duration_secs': 0.456398} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.836873] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 670.838168] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ff6edfe-dc24-4bfd-9d25-900ec001a2c7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.846041] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Waiting for the task: (returnval){ [ 670.846041] env[65522]: value = "task-5113646" [ 670.846041] env[65522]: _type = "Task" [ 670.846041] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.857453] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113646, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.970411] env[65522]: DEBUG nova.compute.manager [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 670.987669] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "refresh_cache-97dc8cac-8241-4912-a3ed-689439ef7ff8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.987889] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquired lock "refresh_cache-97dc8cac-8241-4912-a3ed-689439ef7ff8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 670.988160] env[65522]: DEBUG nova.network.neutron [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 670.999889] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113642, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.040823] env[65522]: WARNING openstack [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.041653] env[65522]: WARNING openstack [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.147750] env[65522]: WARNING neutronclient.v2_0.client [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 671.148829] env[65522]: WARNING openstack [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.149468] env[65522]: WARNING openstack [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.178630] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.232697] env[65522]: DEBUG nova.compute.manager [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Received event network-vif-plugged-97dc7703-dd4a-43dd-8e2d-02f8a31d1a69 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 671.232974] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Acquiring lock "0d52b207-ac69-48ce-b3af-7a83a499886e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.233201] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Lock "0d52b207-ac69-48ce-b3af-7a83a499886e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 671.233392] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Lock "0d52b207-ac69-48ce-b3af-7a83a499886e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 671.233634] env[65522]: DEBUG nova.compute.manager [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] No waiting events found dispatching network-vif-plugged-97dc7703-dd4a-43dd-8e2d-02f8a31d1a69 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 671.233888] env[65522]: WARNING nova.compute.manager [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Received unexpected event network-vif-plugged-97dc7703-dd4a-43dd-8e2d-02f8a31d1a69 for instance with vm_state building and task_state spawning. [ 671.234465] env[65522]: DEBUG nova.compute.manager [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Received event network-changed-97dc7703-dd4a-43dd-8e2d-02f8a31d1a69 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 671.234768] env[65522]: DEBUG nova.compute.manager [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Refreshing instance network info cache due to event network-changed-97dc7703-dd4a-43dd-8e2d-02f8a31d1a69. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 671.235111] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Acquiring lock "refresh_cache-0d52b207-ac69-48ce-b3af-7a83a499886e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.235429] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Acquired lock "refresh_cache-0d52b207-ac69-48ce-b3af-7a83a499886e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 671.235608] env[65522]: DEBUG nova.network.neutron [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Refreshing network info cache for port 97dc7703-dd4a-43dd-8e2d-02f8a31d1a69 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 671.262363] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113645, 'name': CreateVM_Task, 'duration_secs': 0.483021} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.262569] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 671.263103] env[65522]: WARNING neutronclient.v2_0.client [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 671.263477] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.263623] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 671.263942] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 671.264239] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-624ebaac-f411-4467-bfcd-e2c2541c2450 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.268485] env[65522]: DEBUG nova.network.neutron [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Updated VIF entry in instance network info cache for port a7a9fee0-67af-4446-b838-0cd8922465b9. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 671.268906] env[65522]: DEBUG nova.network.neutron [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Updating instance_info_cache with network_info: [{"id": "a7a9fee0-67af-4446-b838-0cd8922465b9", "address": "fa:16:3e:33:16:af", "network": {"id": "29f88cc3-e21a-4a6e-87a2-8624a1be6482", "bridge": "br-int", "label": "tempest-ServersTestJSON-1088062070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f64db4fc2f3412986ef6fd1393e99c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7a9fee0-67", "ovs_interfaceid": "a7a9fee0-67af-4446-b838-0cd8922465b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 671.277946] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 671.277946] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527c0f4b-afd0-11a1-db79-961b04d84e9a" [ 671.277946] env[65522]: _type = "Task" [ 671.277946] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.293017] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527c0f4b-afd0-11a1-db79-961b04d84e9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.358528] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113646, 'name': PowerOnVM_Task} progress is 1%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.494529] env[65522]: WARNING openstack [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.494890] env[65522]: WARNING openstack [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.504617] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.515651] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113642, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.979263} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.515916] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 0d52b207-ac69-48ce-b3af-7a83a499886e/0d52b207-ac69-48ce-b3af-7a83a499886e.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 671.516954] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 671.516954] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a5087b90-5fbf-44a9-af67-548e20888948 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.524963] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 671.524963] env[65522]: value = "task-5113647" [ 671.524963] env[65522]: _type = "Task" [ 671.524963] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.538557] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113647, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.545580] env[65522]: DEBUG nova.network.neutron [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 671.566061] env[65522]: WARNING openstack [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.566061] env[65522]: WARNING openstack [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.720198] env[65522]: WARNING neutronclient.v2_0.client [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 671.721085] env[65522]: WARNING openstack [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.721453] env[65522]: WARNING openstack [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.744493] env[65522]: WARNING neutronclient.v2_0.client [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 671.745173] env[65522]: WARNING openstack [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.745510] env[65522]: WARNING openstack [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.755349] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e9d549-4ed9-416f-b265-e8960ec24a9e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.764710] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02b6edf-86f9-42d4-8fa3-79538ddc3918 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.800572] env[65522]: DEBUG oslo_concurrency.lockutils [req-86e2b713-a849-4e67-aecd-6b7eb17a065f req-2f989d64-8a33-4eeb-9060-912e5170bd88 service nova] Releasing lock "refresh_cache-fdf56fa6-ad18-499e-a171-6242abb293bf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 671.810647] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e48e351-be2c-4088-b7bf-0bae5802689d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.824840] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b4951d8-b66a-4813-a1db-efc4d9db58f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.829802] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527c0f4b-afd0-11a1-db79-961b04d84e9a, 'name': SearchDatastore_Task, 'duration_secs': 0.070271} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.831547] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 671.831776] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 671.832059] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.832205] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 671.834491] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 671.834491] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fb5b6b1b-632c-439a-8f0b-b8197bd63eba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.843940] env[65522]: DEBUG nova.compute.provider_tree [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.847896] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 671.848113] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 671.853488] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b113a04b-1c75-42fc-9feb-7020eccc9b12 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.860924] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 671.860924] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52120498-22fe-c25b-917b-29f1c386cd61" [ 671.860924] env[65522]: _type = "Task" [ 671.860924] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.863605] env[65522]: DEBUG nova.network.neutron [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Updating instance_info_cache with network_info: [{"id": "475e049b-993f-4ab4-aba6-da16d2828c40", "address": "fa:16:3e:28:3d:40", "network": {"id": "6314277f-6ed5-45e5-950d-2683782849aa", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-246048775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "95bca233019f4e2f8bac0dd800392a69", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap475e049b-99", "ovs_interfaceid": "475e049b-993f-4ab4-aba6-da16d2828c40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 671.870891] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113646, 'name': PowerOnVM_Task} progress is 64%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.881889] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.882188] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 671.888574] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52120498-22fe-c25b-917b-29f1c386cd61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.925129] env[65522]: WARNING openstack [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.925535] env[65522]: WARNING openstack [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.941703] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 671.941703] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 672.002600] env[65522]: WARNING neutronclient.v2_0.client [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 672.003827] env[65522]: WARNING openstack [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 672.003827] env[65522]: WARNING openstack [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 672.038219] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113647, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094981} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.038219] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 672.038219] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e50827c-f2f2-4702-ac88-0b7aa3d14a5d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.065989] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Reconfiguring VM instance instance-00000011 to attach disk [datastore1] 0d52b207-ac69-48ce-b3af-7a83a499886e/0d52b207-ac69-48ce-b3af-7a83a499886e.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 672.070588] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ddad64e-240e-4b3e-b64d-a050c3d208ed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.098205] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 672.098205] env[65522]: value = "task-5113648" [ 672.098205] env[65522]: _type = "Task" [ 672.098205] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.108554] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113648, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.139981] env[65522]: DEBUG nova.network.neutron [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Updated VIF entry in instance network info cache for port 97dc7703-dd4a-43dd-8e2d-02f8a31d1a69. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 672.140353] env[65522]: DEBUG nova.network.neutron [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Updating instance_info_cache with network_info: [{"id": "97dc7703-dd4a-43dd-8e2d-02f8a31d1a69", "address": "fa:16:3e:16:68:0f", "network": {"id": "d0ade77e-db62-42ae-9058-c33fcab1d2f9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1207104058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "234a226c969b4965bfa2c411bb4c8d90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97dc7703-dd", "ovs_interfaceid": "97dc7703-dd4a-43dd-8e2d-02f8a31d1a69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 672.347257] env[65522]: DEBUG nova.scheduler.client.report [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 672.362442] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113646, 'name': PowerOnVM_Task} progress is 82%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.372795] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Releasing lock "refresh_cache-97dc8cac-8241-4912-a3ed-689439ef7ff8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 672.373200] env[65522]: DEBUG nova.compute.manager [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Instance network_info: |[{"id": "475e049b-993f-4ab4-aba6-da16d2828c40", "address": "fa:16:3e:28:3d:40", "network": {"id": "6314277f-6ed5-45e5-950d-2683782849aa", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-246048775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "95bca233019f4e2f8bac0dd800392a69", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap475e049b-99", "ovs_interfaceid": "475e049b-993f-4ab4-aba6-da16d2828c40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 672.380693] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:3d:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '136c3499-9ca0-4f85-903d-1f194aa66ed9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '475e049b-993f-4ab4-aba6-da16d2828c40', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 672.395757] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 672.396765] env[65522]: DEBUG nova.compute.manager [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 672.399479] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 672.400872] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-629a671c-d7c4-48b9-a4f0-71abd678d269 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.421759] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52120498-22fe-c25b-917b-29f1c386cd61, 'name': SearchDatastore_Task, 'duration_secs': 0.028368} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.423265] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5567352f-1b70-4708-8787-c95fec594d47 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.433608] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 672.433608] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f50639-59cd-3255-4f2a-a6cf346341d3" [ 672.433608] env[65522]: _type = "Task" [ 672.433608] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.433912] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 672.433912] env[65522]: value = "task-5113649" [ 672.433912] env[65522]: _type = "Task" [ 672.433912] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.459628] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f50639-59cd-3255-4f2a-a6cf346341d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.459628] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113649, 'name': CreateVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.459628] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 672.459980] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 672.460595] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 672.460595] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 672.460706] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 672.460817] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 672.460949] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 672.461219] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 672.613055] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113648, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.642980] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Releasing lock "refresh_cache-0d52b207-ac69-48ce-b3af-7a83a499886e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 672.643270] env[65522]: DEBUG nova.compute.manager [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Received event network-vif-plugged-8cb15276-e492-4096-b65b-f1107c89c902 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 672.643465] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Acquiring lock "a8f9f223-023a-48fc-9186-a360a78e27a9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 672.643668] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Lock "a8f9f223-023a-48fc-9186-a360a78e27a9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 672.643820] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Lock "a8f9f223-023a-48fc-9186-a360a78e27a9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 672.643976] env[65522]: DEBUG nova.compute.manager [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] No waiting events found dispatching network-vif-plugged-8cb15276-e492-4096-b65b-f1107c89c902 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 672.644166] env[65522]: WARNING nova.compute.manager [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Received unexpected event network-vif-plugged-8cb15276-e492-4096-b65b-f1107c89c902 for instance with vm_state building and task_state spawning. [ 672.644343] env[65522]: DEBUG nova.compute.manager [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Received event network-changed-8cb15276-e492-4096-b65b-f1107c89c902 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 672.644508] env[65522]: DEBUG nova.compute.manager [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Refreshing instance network info cache due to event network-changed-8cb15276-e492-4096-b65b-f1107c89c902. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 672.644686] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Acquiring lock "refresh_cache-a8f9f223-023a-48fc-9186-a360a78e27a9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.644816] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Acquired lock "refresh_cache-a8f9f223-023a-48fc-9186-a360a78e27a9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 672.644962] env[65522]: DEBUG nova.network.neutron [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Refreshing network info cache for port 8cb15276-e492-4096-b65b-f1107c89c902 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 672.856407] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.527s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 672.856931] env[65522]: DEBUG nova.compute.manager [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 672.860907] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.634s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 672.861855] env[65522]: INFO nova.compute.claims [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 672.872020] env[65522]: DEBUG oslo_vmware.api [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113646, 'name': PowerOnVM_Task, 'duration_secs': 1.638925} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.872307] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 672.872507] env[65522]: INFO nova.compute.manager [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Took 10.73 seconds to spawn the instance on the hypervisor. [ 672.872682] env[65522]: DEBUG nova.compute.manager [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 672.873529] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfd0fc2-2a5f-467f-88b7-dc4766301140 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.947820] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 672.958627] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f50639-59cd-3255-4f2a-a6cf346341d3, 'name': SearchDatastore_Task, 'duration_secs': 0.032573} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.958848] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113649, 'name': CreateVM_Task, 'duration_secs': 0.472073} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.959130] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 672.960201] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] a8f9f223-023a-48fc-9186-a360a78e27a9/a8f9f223-023a-48fc-9186-a360a78e27a9.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 672.960201] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 672.960201] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f7b4f3d2-1829-41ef-af44-f8b4011b04ac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.963848] env[65522]: WARNING neutronclient.v2_0.client [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 672.964227] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.964410] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 672.964744] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 672.966352] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 672.966475] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f58c20b6-c061-41f4-aa95-9bc1b9daa9b2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.972023] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 672.972023] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52660b9f-ec92-93d8-c7b9-15646f4382b8" [ 672.972023] env[65522]: _type = "Task" [ 672.972023] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.977681] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 672.977681] env[65522]: value = "task-5113650" [ 672.977681] env[65522]: _type = "Task" [ 672.977681] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.986807] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52660b9f-ec92-93d8-c7b9-15646f4382b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.992804] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113650, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.111580] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113648, 'name': ReconfigVM_Task, 'duration_secs': 0.748703} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.111941] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Reconfigured VM instance instance-00000011 to attach disk [datastore1] 0d52b207-ac69-48ce-b3af-7a83a499886e/0d52b207-ac69-48ce-b3af-7a83a499886e.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 673.112527] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0de49a16-b6ee-412c-afec-8bcb1a707218 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.121802] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 673.121802] env[65522]: value = "task-5113651" [ 673.121802] env[65522]: _type = "Task" [ 673.121802] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.135610] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113651, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.149645] env[65522]: WARNING neutronclient.v2_0.client [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 673.149645] env[65522]: WARNING openstack [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.150203] env[65522]: WARNING openstack [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.371441] env[65522]: DEBUG nova.compute.utils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 673.373589] env[65522]: DEBUG nova.compute.manager [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 673.373793] env[65522]: DEBUG nova.network.neutron [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 673.374121] env[65522]: WARNING neutronclient.v2_0.client [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 673.374466] env[65522]: WARNING neutronclient.v2_0.client [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 673.375592] env[65522]: WARNING openstack [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.376287] env[65522]: WARNING openstack [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.405169] env[65522]: INFO nova.compute.manager [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Took 21.51 seconds to build instance. [ 673.463119] env[65522]: DEBUG nova.policy [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '767535aa17a84adeb722a33713b7aa68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b8112ba49034e6fabcb3fbbd46edf41', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 673.501797] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52660b9f-ec92-93d8-c7b9-15646f4382b8, 'name': SearchDatastore_Task, 'duration_secs': 0.035025} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.502383] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113650, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.504979] env[65522]: WARNING openstack [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.504979] env[65522]: WARNING openstack [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.512543] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 673.512815] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 673.513200] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.513485] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 673.513747] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 673.514901] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-040c99be-1f69-4de2-a4c5-e41d3f4efd9b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.531269] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 673.531429] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 673.532864] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfa8f88d-5313-4af6-977d-055ce6b81c81 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.541022] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 673.541022] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52525b01-4f85-100b-cc2e-99a3d84da27f" [ 673.541022] env[65522]: _type = "Task" [ 673.541022] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.553870] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52525b01-4f85-100b-cc2e-99a3d84da27f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.641341] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113651, 'name': Rename_Task, 'duration_secs': 0.289358} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.641776] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 673.646875] env[65522]: DEBUG nova.compute.manager [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Received event network-vif-plugged-475e049b-993f-4ab4-aba6-da16d2828c40 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 673.647220] env[65522]: DEBUG oslo_concurrency.lockutils [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] Acquiring lock "97dc8cac-8241-4912-a3ed-689439ef7ff8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 673.647425] env[65522]: DEBUG oslo_concurrency.lockutils [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] Lock "97dc8cac-8241-4912-a3ed-689439ef7ff8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 673.647591] env[65522]: DEBUG oslo_concurrency.lockutils [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] Lock "97dc8cac-8241-4912-a3ed-689439ef7ff8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 673.647774] env[65522]: DEBUG nova.compute.manager [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] No waiting events found dispatching network-vif-plugged-475e049b-993f-4ab4-aba6-da16d2828c40 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 673.647936] env[65522]: WARNING nova.compute.manager [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Received unexpected event network-vif-plugged-475e049b-993f-4ab4-aba6-da16d2828c40 for instance with vm_state building and task_state spawning. [ 673.648384] env[65522]: DEBUG nova.compute.manager [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Received event network-changed-475e049b-993f-4ab4-aba6-da16d2828c40 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 673.648384] env[65522]: DEBUG nova.compute.manager [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Refreshing instance network info cache due to event network-changed-475e049b-993f-4ab4-aba6-da16d2828c40. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 673.649216] env[65522]: DEBUG oslo_concurrency.lockutils [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] Acquiring lock "refresh_cache-97dc8cac-8241-4912-a3ed-689439ef7ff8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.649216] env[65522]: DEBUG oslo_concurrency.lockutils [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] Acquired lock "refresh_cache-97dc8cac-8241-4912-a3ed-689439ef7ff8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 673.649216] env[65522]: DEBUG nova.network.neutron [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Refreshing network info cache for port 475e049b-993f-4ab4-aba6-da16d2828c40 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 673.650632] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-311109b4-356a-4b53-b8fa-ec40e1c5724e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.659701] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 673.659701] env[65522]: value = "task-5113652" [ 673.659701] env[65522]: _type = "Task" [ 673.659701] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.677242] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113652, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.886028] env[65522]: DEBUG nova.compute.manager [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 673.910119] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f35325da-cb84-48ed-ace6-79ee7bac7e19 tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Lock "fdf56fa6-ad18-499e-a171-6242abb293bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.025s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 673.951463] env[65522]: DEBUG nova.network.neutron [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Successfully created port: ede8edfa-842b-482d-8e7b-34bde13a02f8 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 673.989895] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113650, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.848758} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.990172] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] a8f9f223-023a-48fc-9186-a360a78e27a9/a8f9f223-023a-48fc-9186-a360a78e27a9.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 673.990392] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 673.990646] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f94506e-e4ca-48ee-8b65-c4bebe5b7d6a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.001418] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 674.001418] env[65522]: value = "task-5113653" [ 674.001418] env[65522]: _type = "Task" [ 674.001418] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.011760] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113653, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.055306] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52525b01-4f85-100b-cc2e-99a3d84da27f, 'name': SearchDatastore_Task, 'duration_secs': 0.029226} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.060878] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10be92ae-f605-44aa-a28c-1f383f5b87ca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.066839] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 674.066839] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522a0fff-10cd-6402-22b1-dad0d3bcff2a" [ 674.066839] env[65522]: _type = "Task" [ 674.066839] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.082867] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522a0fff-10cd-6402-22b1-dad0d3bcff2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.153984] env[65522]: WARNING neutronclient.v2_0.client [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 674.154705] env[65522]: WARNING openstack [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.155053] env[65522]: WARNING openstack [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.176556] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113652, 'name': PowerOnVM_Task} progress is 80%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.243959] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db694176-de49-40d1-8a71-45864186c6e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.253109] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb7849c-0325-4e73-a437-93aa3c9a3610 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.289853] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a815168-cfe2-41a2-8324-26d7ae6eaa48 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.300649] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b74c50cf-bd91-459f-be8b-d0cf60645ccb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.316106] env[65522]: DEBUG nova.compute.provider_tree [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.513398] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113653, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.328314} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.513607] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 674.514535] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83a8f63-0a44-42e5-92cf-c1cfd4105e56 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.538217] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Reconfiguring VM instance instance-00000012 to attach disk [datastore1] a8f9f223-023a-48fc-9186-a360a78e27a9/a8f9f223-023a-48fc-9186-a360a78e27a9.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 674.538549] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8d3aad6-65f8-47a5-b322-4b3a3631b267 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.561709] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 674.561709] env[65522]: value = "task-5113654" [ 674.561709] env[65522]: _type = "Task" [ 674.561709] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.573810] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113654, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.581216] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522a0fff-10cd-6402-22b1-dad0d3bcff2a, 'name': SearchDatastore_Task, 'duration_secs': 0.040454} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.581216] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 674.581216] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 97dc8cac-8241-4912-a3ed-689439ef7ff8/97dc8cac-8241-4912-a3ed-689439ef7ff8.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 674.582779] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a595739-554e-4247-8795-00ca3ec3c451 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.589027] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 674.589027] env[65522]: value = "task-5113655" [ 674.589027] env[65522]: _type = "Task" [ 674.589027] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.598202] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113655, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.675904] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113652, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.679148] env[65522]: WARNING neutronclient.v2_0.client [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 674.679148] env[65522]: WARNING openstack [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.679148] env[65522]: WARNING openstack [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.820651] env[65522]: DEBUG nova.scheduler.client.report [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 674.899351] env[65522]: DEBUG nova.compute.manager [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 674.930377] env[65522]: DEBUG nova.virt.hardware [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 674.930377] env[65522]: DEBUG nova.virt.hardware [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 674.930989] env[65522]: DEBUG nova.virt.hardware [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 674.931385] env[65522]: DEBUG nova.virt.hardware [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 674.931664] env[65522]: DEBUG nova.virt.hardware [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 674.931905] env[65522]: DEBUG nova.virt.hardware [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 674.932236] env[65522]: DEBUG nova.virt.hardware [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 674.932490] env[65522]: DEBUG nova.virt.hardware [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 674.932782] env[65522]: DEBUG nova.virt.hardware [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 674.933068] env[65522]: DEBUG nova.virt.hardware [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 674.933361] env[65522]: DEBUG nova.virt.hardware [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 674.935281] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91cf10f9-2dcc-4835-8a2f-a6a1d14d7cbd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.945446] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a8eda7-5cde-49b6-a81b-110a9c393e04 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.075131] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113654, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.103386] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113655, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.177549] env[65522]: DEBUG oslo_vmware.api [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113652, 'name': PowerOnVM_Task, 'duration_secs': 1.152142} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.177887] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 675.178204] env[65522]: INFO nova.compute.manager [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Took 10.41 seconds to spawn the instance on the hypervisor. [ 675.178470] env[65522]: DEBUG nova.compute.manager [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 675.179397] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-266b6d8c-e126-4ee9-b818-afb3e865b397 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.328194] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 675.328803] env[65522]: DEBUG nova.compute.manager [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 675.331910] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.863s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 675.334150] env[65522]: INFO nova.compute.claims [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 675.574867] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113654, 'name': ReconfigVM_Task, 'duration_secs': 0.525666} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.575336] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Reconfigured VM instance instance-00000012 to attach disk [datastore1] a8f9f223-023a-48fc-9186-a360a78e27a9/a8f9f223-023a-48fc-9186-a360a78e27a9.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 675.576096] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-33175b63-7e60-43eb-8285-b5527bfc164a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.584572] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 675.584572] env[65522]: value = "task-5113656" [ 675.584572] env[65522]: _type = "Task" [ 675.584572] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.602224] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113656, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.610405] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113655, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.872171} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.610733] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 97dc8cac-8241-4912-a3ed-689439ef7ff8/97dc8cac-8241-4912-a3ed-689439ef7ff8.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 675.610996] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 675.611352] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-742b9731-14c3-4481-aa94-7278cf58b915 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.621825] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 675.621825] env[65522]: value = "task-5113657" [ 675.621825] env[65522]: _type = "Task" [ 675.621825] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.632903] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113657, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.699891] env[65522]: INFO nova.compute.manager [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Took 21.21 seconds to build instance. [ 675.840750] env[65522]: DEBUG nova.compute.utils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 675.844952] env[65522]: DEBUG nova.compute.manager [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 675.845174] env[65522]: DEBUG nova.network.neutron [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 675.845496] env[65522]: WARNING neutronclient.v2_0.client [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 675.845816] env[65522]: WARNING neutronclient.v2_0.client [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 675.846381] env[65522]: WARNING openstack [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 675.847034] env[65522]: WARNING openstack [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.981789] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquiring lock "dbf29bd1-780c-4756-a203-4fad05a9d3b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 675.982025] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lock "dbf29bd1-780c-4756-a203-4fad05a9d3b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 676.240153] env[65522]: DEBUG nova.network.neutron [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Successfully updated port: ede8edfa-842b-482d-8e7b-34bde13a02f8 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 676.240153] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113656, 'name': Rename_Task, 'duration_secs': 0.355268} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.240153] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 676.240153] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d35850a-5a49-4b34-a24e-eb9ed0ca2ead {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.240153] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 676.240153] env[65522]: value = "task-5113658" [ 676.240153] env[65522]: _type = "Task" [ 676.240153] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.242052] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113658, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.242052] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113657, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.211173} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.242052] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 676.242052] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17a8c7e5-5817-48d6-b491-5549b89822c8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.242052] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Reconfiguring VM instance instance-00000013 to attach disk [datastore2] 97dc8cac-8241-4912-a3ed-689439ef7ff8/97dc8cac-8241-4912-a3ed-689439ef7ff8.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 676.242643] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-302eb0ec-1b42-4b6c-9f9e-12b4d4f5ead5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.242643] env[65522]: DEBUG oslo_concurrency.lockutils [None req-52317b87-48b0-4004-a866-bd08d2916c8b tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "0d52b207-ac69-48ce-b3af-7a83a499886e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.726s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 676.242643] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 676.242643] env[65522]: value = "task-5113659" [ 676.242643] env[65522]: _type = "Task" [ 676.242643] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.242643] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113659, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.348138] env[65522]: DEBUG nova.compute.manager [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 676.485403] env[65522]: DEBUG nova.compute.manager [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 676.557218] env[65522]: DEBUG nova.policy [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24bc4117f7884e7d8a9333607b1276e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '95bca233019f4e2f8bac0dd800392a69', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 676.592735] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "refresh_cache-ca964440-5375-4aff-8b45-96fbe829dd16" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.592926] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "refresh_cache-ca964440-5375-4aff-8b45-96fbe829dd16" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 676.593117] env[65522]: DEBUG nova.network.neutron [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 676.596400] env[65522]: DEBUG nova.network.neutron [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Updated VIF entry in instance network info cache for port 8cb15276-e492-4096-b65b-f1107c89c902. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 676.596729] env[65522]: DEBUG nova.network.neutron [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Updating instance_info_cache with network_info: [{"id": "8cb15276-e492-4096-b65b-f1107c89c902", "address": "fa:16:3e:05:fb:2d", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cb15276-e4", "ovs_interfaceid": "8cb15276-e492-4096-b65b-f1107c89c902", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 676.617544] env[65522]: WARNING openstack [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.617895] env[65522]: WARNING openstack [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 676.636056] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113658, 'name': PowerOnVM_Task} progress is 1%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.720940] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113659, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.813514] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa55f4a-ff26-4ea1-83b2-e15d2b73c072 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.823099] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386c6fc3-f7c5-49de-8768-1beff0c5f04f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.863724] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a4b16ad-b566-4fe6-9e4f-13825c6fad13 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.874329] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25ec7e4-2e79-4f90-88bd-0839e6a21939 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.891573] env[65522]: DEBUG nova.compute.provider_tree [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.028542] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 677.099895] env[65522]: WARNING openstack [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.100283] env[65522]: WARNING openstack [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.113140] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdfbbf7e-0817-4b2e-8531-33280457dd33 req-487d63d1-4c1d-40fd-82c9-d0462ec4f310 service nova] Releasing lock "refresh_cache-a8f9f223-023a-48fc-9186-a360a78e27a9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 677.127978] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113658, 'name': PowerOnVM_Task} progress is 1%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.219220] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113659, 'name': ReconfigVM_Task, 'duration_secs': 0.71532} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.219220] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Reconfigured VM instance instance-00000013 to attach disk [datastore2] 97dc8cac-8241-4912-a3ed-689439ef7ff8/97dc8cac-8241-4912-a3ed-689439ef7ff8.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 677.219493] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c443bdd-723d-43d3-905d-05886f27d0f5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.226886] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 677.226886] env[65522]: value = "task-5113660" [ 677.226886] env[65522]: _type = "Task" [ 677.226886] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.237163] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113660, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.338283] env[65522]: DEBUG nova.network.neutron [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 677.362130] env[65522]: DEBUG nova.network.neutron [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Successfully created port: ff1e8448-0d72-4738-ae2b-9b543dbe9f5e {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 677.365602] env[65522]: DEBUG nova.compute.manager [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 677.395543] env[65522]: DEBUG nova.scheduler.client.report [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 677.402389] env[65522]: WARNING neutronclient.v2_0.client [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 677.403160] env[65522]: WARNING openstack [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.403550] env[65522]: WARNING openstack [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.422655] env[65522]: DEBUG nova.virt.hardware [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 677.422935] env[65522]: DEBUG nova.virt.hardware [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 677.423142] env[65522]: DEBUG nova.virt.hardware [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 677.423323] env[65522]: DEBUG nova.virt.hardware [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 677.423479] env[65522]: DEBUG nova.virt.hardware [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 677.423623] env[65522]: DEBUG nova.virt.hardware [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 677.423829] env[65522]: DEBUG nova.virt.hardware [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.424199] env[65522]: DEBUG nova.virt.hardware [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 677.424199] env[65522]: DEBUG nova.virt.hardware [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 677.424429] env[65522]: DEBUG nova.virt.hardware [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 677.424510] env[65522]: DEBUG nova.virt.hardware [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 677.425799] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ffd1a1-7dc3-4c13-a03c-86f54198b598 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.437633] env[65522]: WARNING openstack [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.437938] env[65522]: WARNING openstack [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.446208] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c0c1d6-1532-419d-af4b-75ad1b80a94c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.627689] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113658, 'name': PowerOnVM_Task} progress is 64%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.745145] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113660, 'name': Rename_Task, 'duration_secs': 0.224881} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.745419] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 677.745672] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-843dc9cc-9a13-43fa-a9ff-dc8bb2162a77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.754686] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 677.754686] env[65522]: value = "task-5113661" [ 677.754686] env[65522]: _type = "Task" [ 677.754686] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.765386] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113661, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.880480] env[65522]: DEBUG nova.network.neutron [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Updated VIF entry in instance network info cache for port 475e049b-993f-4ab4-aba6-da16d2828c40. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 677.880895] env[65522]: DEBUG nova.network.neutron [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Updating instance_info_cache with network_info: [{"id": "475e049b-993f-4ab4-aba6-da16d2828c40", "address": "fa:16:3e:28:3d:40", "network": {"id": "6314277f-6ed5-45e5-950d-2683782849aa", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-246048775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95bca233019f4e2f8bac0dd800392a69", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap475e049b-99", "ovs_interfaceid": "475e049b-993f-4ab4-aba6-da16d2828c40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 677.895828] env[65522]: WARNING neutronclient.v2_0.client [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 677.896536] env[65522]: WARNING openstack [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.896887] env[65522]: WARNING openstack [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.905316] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 677.906307] env[65522]: DEBUG nova.compute.manager [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 677.908683] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.262s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 677.910137] env[65522]: INFO nova.compute.claims [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 678.057931] env[65522]: DEBUG nova.network.neutron [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Updating instance_info_cache with network_info: [{"id": "ede8edfa-842b-482d-8e7b-34bde13a02f8", "address": "fa:16:3e:f6:50:03", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapede8edfa-84", "ovs_interfaceid": "ede8edfa-842b-482d-8e7b-34bde13a02f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 678.125679] env[65522]: DEBUG oslo_vmware.api [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113658, 'name': PowerOnVM_Task, 'duration_secs': 1.747676} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.126161] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 678.127658] env[65522]: INFO nova.compute.manager [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Took 11.28 seconds to spawn the instance on the hypervisor. [ 678.128048] env[65522]: DEBUG nova.compute.manager [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 678.130971] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e321fbd-461c-422c-8262-b7517e42c312 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.267362] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113661, 'name': PowerOnVM_Task} progress is 68%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.385142] env[65522]: DEBUG oslo_concurrency.lockutils [req-390289ed-2c19-41fb-bf54-d0cc96b97a14 req-db92f030-885a-49b6-b843-24580dfdaeb3 service nova] Releasing lock "refresh_cache-97dc8cac-8241-4912-a3ed-689439ef7ff8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 678.417265] env[65522]: DEBUG nova.compute.utils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 678.419524] env[65522]: DEBUG nova.compute.manager [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 678.419524] env[65522]: DEBUG nova.network.neutron [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 678.419728] env[65522]: WARNING neutronclient.v2_0.client [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 678.420917] env[65522]: WARNING neutronclient.v2_0.client [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 678.420917] env[65522]: WARNING openstack [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 678.420917] env[65522]: WARNING openstack [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 678.566018] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "refresh_cache-ca964440-5375-4aff-8b45-96fbe829dd16" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 678.566018] env[65522]: DEBUG nova.compute.manager [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Instance network_info: |[{"id": "ede8edfa-842b-482d-8e7b-34bde13a02f8", "address": "fa:16:3e:f6:50:03", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapede8edfa-84", "ovs_interfaceid": "ede8edfa-842b-482d-8e7b-34bde13a02f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 678.566299] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:50:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4b033f4d-2e92-4702-add6-410a29d3f251', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ede8edfa-842b-482d-8e7b-34bde13a02f8', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 678.575434] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Creating folder: Project (9b8112ba49034e6fabcb3fbbd46edf41). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 678.575966] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c46b91e0-e222-487e-8fab-a4b9d506b723 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.589349] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Created folder: Project (9b8112ba49034e6fabcb3fbbd46edf41) in parent group-v994660. [ 678.589875] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Creating folder: Instances. Parent ref: group-v994717. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 678.590332] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1cb0ea39-b752-4ad5-946d-48ed11924c99 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.602653] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Created folder: Instances in parent group-v994717. [ 678.603202] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 678.604538] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 678.604538] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1ceb3878-b990-4903-90b2-78a08df18929 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.623938] env[65522]: DEBUG nova.policy [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '767535aa17a84adeb722a33713b7aa68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b8112ba49034e6fabcb3fbbd46edf41', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 678.632514] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 678.632514] env[65522]: value = "task-5113664" [ 678.632514] env[65522]: _type = "Task" [ 678.632514] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.642930] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113664, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.652129] env[65522]: INFO nova.compute.manager [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Took 21.75 seconds to build instance. [ 678.769154] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113661, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.929118] env[65522]: DEBUG nova.compute.manager [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 679.145504] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113664, 'name': CreateVM_Task, 'duration_secs': 0.447529} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.145691] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 679.146231] env[65522]: WARNING neutronclient.v2_0.client [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 679.146607] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.146765] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.147104] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 679.147382] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-341eabab-d535-4479-b8a9-117d5baed57d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.154121] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b8c9299-5f37-41ec-889f-397acac4670c tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "a8f9f223-023a-48fc-9186-a360a78e27a9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.263s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 679.157019] env[65522]: DEBUG nova.network.neutron [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Successfully updated port: ff1e8448-0d72-4738-ae2b-9b543dbe9f5e {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 679.159661] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 679.159661] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528ac91a-1ace-57a4-e351-919739c5e1ee" [ 679.159661] env[65522]: _type = "Task" [ 679.159661] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.169325] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528ac91a-1ace-57a4-e351-919739c5e1ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.271171] env[65522]: DEBUG oslo_vmware.api [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113661, 'name': PowerOnVM_Task, 'duration_secs': 1.146601} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.271593] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 679.271862] env[65522]: INFO nova.compute.manager [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Took 9.90 seconds to spawn the instance on the hypervisor. [ 679.272059] env[65522]: DEBUG nova.compute.manager [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 679.272877] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f724b58c-29f7-456c-9c98-71f7aeb46f4e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.278177] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42909c3-d710-458d-8053-c4aaf0acd609 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.293900] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95881094-b59d-43f6-b7b4-917cdb224f93 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.334029] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f80ee2-247a-47ae-a3ec-7494566fd19e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.343083] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14cc4f00-4264-4db8-879f-64b7bb1edaea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.358500] env[65522]: DEBUG nova.compute.provider_tree [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.477331] env[65522]: DEBUG nova.network.neutron [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Successfully created port: 60d5d352-4fed-4b0f-9efc-ce13e2066444 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 679.666436] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "refresh_cache-cd82f2fd-d657-4ece-8e90-9a47072f5546" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.666436] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquired lock "refresh_cache-cd82f2fd-d657-4ece-8e90-9a47072f5546" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.666436] env[65522]: DEBUG nova.network.neutron [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 679.674956] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528ac91a-1ace-57a4-e351-919739c5e1ee, 'name': SearchDatastore_Task, 'duration_secs': 0.014116} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.675714] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 679.675959] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 679.676205] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.676385] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.676511] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 679.676893] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d8eded2-d71e-4620-ab97-8cc791edc0e0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.688928] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 679.688928] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 679.689943] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-334fb8fd-b819-4f43-bd6e-ebf07c4660e4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.702022] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 679.702022] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528ad80e-5dd1-51cb-e177-9b521fb3e51d" [ 679.702022] env[65522]: _type = "Task" [ 679.702022] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.711463] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528ad80e-5dd1-51cb-e177-9b521fb3e51d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.798521] env[65522]: INFO nova.compute.manager [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Took 22.81 seconds to build instance. [ 679.861884] env[65522]: DEBUG nova.scheduler.client.report [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 679.939678] env[65522]: DEBUG nova.compute.manager [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 679.964873] env[65522]: DEBUG nova.virt.hardware [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 679.965120] env[65522]: DEBUG nova.virt.hardware [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 679.965274] env[65522]: DEBUG nova.virt.hardware [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 679.965448] env[65522]: DEBUG nova.virt.hardware [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 679.965588] env[65522]: DEBUG nova.virt.hardware [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 679.965729] env[65522]: DEBUG nova.virt.hardware [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 679.965925] env[65522]: DEBUG nova.virt.hardware [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 679.966155] env[65522]: DEBUG nova.virt.hardware [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 679.966336] env[65522]: DEBUG nova.virt.hardware [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 679.966495] env[65522]: DEBUG nova.virt.hardware [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 679.966664] env[65522]: DEBUG nova.virt.hardware [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 679.967561] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0aac0a-5d4b-4519-ad04-328659565698 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.976633] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2d11c2-76ee-468a-91ed-dd280e31198e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.177961] env[65522]: WARNING openstack [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.177961] env[65522]: WARNING openstack [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.213302] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528ad80e-5dd1-51cb-e177-9b521fb3e51d, 'name': SearchDatastore_Task, 'duration_secs': 0.011817} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.214121] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8655a86b-6cd9-4cff-aea1-5d8aa1e07564 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.221792] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 680.221792] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52026a2f-81f5-5fe2-bc26-faf2ae73ab15" [ 680.221792] env[65522]: _type = "Task" [ 680.221792] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.231606] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52026a2f-81f5-5fe2-bc26-faf2ae73ab15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.300453] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9cdff5b9-008d-4003-b43f-3556dbb769cb tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "97dc8cac-8241-4912-a3ed-689439ef7ff8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.327s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 680.367927] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 680.368547] env[65522]: DEBUG nova.compute.manager [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 680.371837] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.798s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.372102] env[65522]: DEBUG nova.objects.instance [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lazy-loading 'resources' on Instance uuid c93b52c3-1143-44d7-8c10-ac58bc25f541 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 680.418148] env[65522]: DEBUG nova.network.neutron [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 680.432171] env[65522]: DEBUG nova.compute.manager [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Received event network-vif-deleted-573a1035-e00a-4cd2-a067-934487b65b60 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 680.432171] env[65522]: DEBUG nova.compute.manager [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Received event network-vif-deleted-366aac9d-7922-4057-b141-5f4d54f4e2ab {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 680.432171] env[65522]: DEBUG nova.compute.manager [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Received event network-changed-7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 680.432171] env[65522]: DEBUG nova.compute.manager [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Refreshing instance network info cache due to event network-changed-7d49f270-6cbd-4767-91fc-f899c46564c9. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 680.432171] env[65522]: DEBUG oslo_concurrency.lockutils [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] Acquiring lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.432510] env[65522]: DEBUG oslo_concurrency.lockutils [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] Acquired lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 680.432510] env[65522]: DEBUG nova.network.neutron [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Refreshing network info cache for port 7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 680.902461] env[65522]: DEBUG nova.compute.utils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 680.913416] env[65522]: DEBUG nova.compute.manager [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 680.914727] env[65522]: DEBUG nova.network.neutron [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 680.917574] env[65522]: WARNING neutronclient.v2_0.client [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.917574] env[65522]: WARNING neutronclient.v2_0.client [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.917574] env[65522]: WARNING openstack [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.917574] env[65522]: WARNING openstack [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.925173] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52026a2f-81f5-5fe2-bc26-faf2ae73ab15, 'name': SearchDatastore_Task, 'duration_secs': 0.010402} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.925903] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.926209] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] ca964440-5375-4aff-8b45-96fbe829dd16/ca964440-5375-4aff-8b45-96fbe829dd16.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 680.926515] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e631573f-50ce-4d96-9a40-a1d1e57a0662 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.935697] env[65522]: WARNING neutronclient.v2_0.client [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.936595] env[65522]: WARNING openstack [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.936595] env[65522]: WARNING openstack [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.948333] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 680.948333] env[65522]: value = "task-5113665" [ 680.948333] env[65522]: _type = "Task" [ 680.948333] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.958870] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113665, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.259108] env[65522]: DEBUG nova.network.neutron [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Successfully updated port: 60d5d352-4fed-4b0f-9efc-ce13e2066444 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 681.269114] env[65522]: WARNING openstack [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 681.269610] env[65522]: WARNING openstack [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 681.335947] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454cf4f8-d3a6-4e93-a4e1-340ad2123227 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.348238] env[65522]: DEBUG nova.policy [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea59d76584dd491a9b8ebeae2a206b54', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb9a9e5da1194ab7927f72d21d1484b0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 681.356152] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46702eb8-48d4-4871-9904-f9609e0394d2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.403267] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c88f947-b1ea-4802-bbe8-751aea805e93 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.413378] env[65522]: DEBUG nova.compute.manager [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 681.419132] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e757ce-4d28-4f4b-99d0-d2c1d19d26b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.449279] env[65522]: DEBUG nova.compute.provider_tree [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.459317] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113665, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.512832] env[65522]: WARNING neutronclient.v2_0.client [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 681.514848] env[65522]: WARNING openstack [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 681.515207] env[65522]: WARNING openstack [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 681.595792] env[65522]: WARNING openstack [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 681.596246] env[65522]: WARNING openstack [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 681.764624] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "refresh_cache-87a3c63d-794d-44ab-bad6-65c323d72ae7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.764792] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "refresh_cache-87a3c63d-794d-44ab-bad6-65c323d72ae7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 681.764956] env[65522]: DEBUG nova.network.neutron [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 681.778766] env[65522]: DEBUG nova.network.neutron [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Successfully created port: 3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 681.793245] env[65522]: DEBUG nova.network.neutron [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Updating instance_info_cache with network_info: [{"id": "ff1e8448-0d72-4738-ae2b-9b543dbe9f5e", "address": "fa:16:3e:b2:40:68", "network": {"id": "6314277f-6ed5-45e5-950d-2683782849aa", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-246048775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95bca233019f4e2f8bac0dd800392a69", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff1e8448-0d", "ovs_interfaceid": "ff1e8448-0d72-4738-ae2b-9b543dbe9f5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 681.823197] env[65522]: WARNING neutronclient.v2_0.client [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 681.823972] env[65522]: WARNING openstack [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 681.824359] env[65522]: WARNING openstack [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 681.950999] env[65522]: DEBUG nova.scheduler.client.report [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 681.969511] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113665, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539593} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.969824] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] ca964440-5375-4aff-8b45-96fbe829dd16/ca964440-5375-4aff-8b45-96fbe829dd16.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 681.970969] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 681.971086] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d040602a-374b-43f6-b87d-a68ba855780f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.988514] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 681.988514] env[65522]: value = "task-5113666" [ 681.988514] env[65522]: _type = "Task" [ 681.988514] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.997924] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113666, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.113860] env[65522]: DEBUG nova.network.neutron [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Updated VIF entry in instance network info cache for port 7d49f270-6cbd-4767-91fc-f899c46564c9. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 682.114142] env[65522]: DEBUG nova.network.neutron [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Updating instance_info_cache with network_info: [{"id": "7d49f270-6cbd-4767-91fc-f899c46564c9", "address": "fa:16:3e:af:04:9f", "network": {"id": "6a60138b-7f03-4817-ac9a-1facdc1cc0cc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1344754077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4fdbd0ed78c41a591ccc0a587313ec8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f85835c8-5d0c-4b2f-97c4-6c4006580f79", "external-id": "nsx-vlan-transportzone-245", "segmentation_id": 245, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d49f270-6c", "ovs_interfaceid": "7d49f270-6cbd-4767-91fc-f899c46564c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 682.226258] env[65522]: DEBUG nova.compute.manager [req-0582cb0a-8956-4f4f-9db4-bf9f14bb7ed0 req-b3761d5e-9c00-47bb-b7ff-053f8ddda1c2 service nova] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Received event network-vif-plugged-ede8edfa-842b-482d-8e7b-34bde13a02f8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 682.226518] env[65522]: DEBUG oslo_concurrency.lockutils [req-0582cb0a-8956-4f4f-9db4-bf9f14bb7ed0 req-b3761d5e-9c00-47bb-b7ff-053f8ddda1c2 service nova] Acquiring lock "ca964440-5375-4aff-8b45-96fbe829dd16-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 682.228841] env[65522]: DEBUG oslo_concurrency.lockutils [req-0582cb0a-8956-4f4f-9db4-bf9f14bb7ed0 req-b3761d5e-9c00-47bb-b7ff-053f8ddda1c2 service nova] Lock "ca964440-5375-4aff-8b45-96fbe829dd16-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 682.228841] env[65522]: DEBUG oslo_concurrency.lockutils [req-0582cb0a-8956-4f4f-9db4-bf9f14bb7ed0 req-b3761d5e-9c00-47bb-b7ff-053f8ddda1c2 service nova] Lock "ca964440-5375-4aff-8b45-96fbe829dd16-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 682.228841] env[65522]: DEBUG nova.compute.manager [req-0582cb0a-8956-4f4f-9db4-bf9f14bb7ed0 req-b3761d5e-9c00-47bb-b7ff-053f8ddda1c2 service nova] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] No waiting events found dispatching network-vif-plugged-ede8edfa-842b-482d-8e7b-34bde13a02f8 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 682.228841] env[65522]: WARNING nova.compute.manager [req-0582cb0a-8956-4f4f-9db4-bf9f14bb7ed0 req-b3761d5e-9c00-47bb-b7ff-053f8ddda1c2 service nova] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Received unexpected event network-vif-plugged-ede8edfa-842b-482d-8e7b-34bde13a02f8 for instance with vm_state building and task_state spawning. [ 682.269986] env[65522]: WARNING openstack [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 682.270440] env[65522]: WARNING openstack [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 682.298024] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Releasing lock "refresh_cache-cd82f2fd-d657-4ece-8e90-9a47072f5546" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 682.298024] env[65522]: DEBUG nova.compute.manager [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Instance network_info: |[{"id": "ff1e8448-0d72-4738-ae2b-9b543dbe9f5e", "address": "fa:16:3e:b2:40:68", "network": {"id": "6314277f-6ed5-45e5-950d-2683782849aa", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-246048775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95bca233019f4e2f8bac0dd800392a69", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff1e8448-0d", "ovs_interfaceid": "ff1e8448-0d72-4738-ae2b-9b543dbe9f5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 682.298490] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b2:40:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '136c3499-9ca0-4f85-903d-1f194aa66ed9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff1e8448-0d72-4738-ae2b-9b543dbe9f5e', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 682.307819] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 682.307911] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 682.308151] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ebea2f2d-82e4-4330-86d0-c1b990d359ba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.332136] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 682.332136] env[65522]: value = "task-5113667" [ 682.332136] env[65522]: _type = "Task" [ 682.332136] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.341594] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113667, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.353497] env[65522]: DEBUG nova.network.neutron [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 682.396521] env[65522]: WARNING openstack [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 682.396521] env[65522]: WARNING openstack [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 682.435377] env[65522]: DEBUG nova.compute.manager [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 682.462603] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.091s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 682.466347] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.288s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 682.466347] env[65522]: DEBUG nova.objects.instance [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Lazy-loading 'resources' on Instance uuid 3f749f68-879f-44f3-8477-7ecea947078c {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 682.473497] env[65522]: DEBUG nova.virt.hardware [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 682.473497] env[65522]: DEBUG nova.virt.hardware [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 682.473497] env[65522]: DEBUG nova.virt.hardware [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 682.473939] env[65522]: DEBUG nova.virt.hardware [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 682.473939] env[65522]: DEBUG nova.virt.hardware [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 682.473939] env[65522]: DEBUG nova.virt.hardware [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 682.473939] env[65522]: DEBUG nova.virt.hardware [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 682.473939] env[65522]: DEBUG nova.virt.hardware [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 682.474391] env[65522]: DEBUG nova.virt.hardware [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 682.474872] env[65522]: DEBUG nova.virt.hardware [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 682.475304] env[65522]: DEBUG nova.virt.hardware [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 682.476533] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d58422-5214-4089-86c4-40d5dc005a33 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.487846] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3c6500-2fb8-4bec-b6ba-8f1b0b489f93 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.498988] env[65522]: INFO nova.scheduler.client.report [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleted allocations for instance c93b52c3-1143-44d7-8c10-ac58bc25f541 [ 682.516339] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113666, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083969} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.519879] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 682.521154] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52cf73e-bf8b-469b-b3bd-a65f6cb6d337 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.547475] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Reconfiguring VM instance instance-00000014 to attach disk [datastore2] ca964440-5375-4aff-8b45-96fbe829dd16/ca964440-5375-4aff-8b45-96fbe829dd16.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 682.547890] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ccdd07b-f380-4ba8-8c4e-a0fae8826328 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.578377] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 682.578377] env[65522]: value = "task-5113668" [ 682.578377] env[65522]: _type = "Task" [ 682.578377] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.591162] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113668, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.617895] env[65522]: DEBUG oslo_concurrency.lockutils [req-be08f648-9d37-464b-90b3-3958e21336e5 req-b4c5922a-7bbb-41fb-a789-84ead13be95e service nova] Releasing lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 682.848736] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113667, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.850256] env[65522]: WARNING neutronclient.v2_0.client [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 682.850862] env[65522]: WARNING openstack [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 682.852109] env[65522]: WARNING openstack [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 683.025790] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3acba54f-5262-4422-9278-c0a4b268f3e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "c93b52c3-1143-44d7-8c10-ac58bc25f541" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.369s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 683.091859] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113668, 'name': ReconfigVM_Task, 'duration_secs': 0.429294} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.092778] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Reconfigured VM instance instance-00000014 to attach disk [datastore2] ca964440-5375-4aff-8b45-96fbe829dd16/ca964440-5375-4aff-8b45-96fbe829dd16.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 683.093013] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e4d780f4-dd50-417b-abbb-2a8b264165d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.101136] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 683.101136] env[65522]: value = "task-5113669" [ 683.101136] env[65522]: _type = "Task" [ 683.101136] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.110637] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113669, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.310226] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab8e6ae-009f-457d-82a3-5c89234ce5e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.320970] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d5e47ac-4a65-48d1-ad4b-886231cc1490 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.355885] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827499d8-95d9-41f2-895c-20cce59511a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.367821] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64b814f7-ee28-49d1-ad2c-24ffe1781e6c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.372064] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113667, 'name': CreateVM_Task, 'duration_secs': 0.812376} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.372205] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 683.373065] env[65522]: WARNING neutronclient.v2_0.client [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 683.373445] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.373596] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 683.373905] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 683.374182] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63843248-6616-44e7-8e24-9e68cbb154e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.383977] env[65522]: DEBUG nova.compute.provider_tree [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.389695] env[65522]: DEBUG nova.network.neutron [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Successfully updated port: 3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 683.396482] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 683.396482] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a58136-abfd-d1b0-7082-06cdf5143e18" [ 683.396482] env[65522]: _type = "Task" [ 683.396482] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.408572] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a58136-abfd-d1b0-7082-06cdf5143e18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.614091] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113669, 'name': Rename_Task, 'duration_secs': 0.160124} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.615129] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 683.615607] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb0554f5-de45-439c-80f6-04f7662ee7d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.626696] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 683.626696] env[65522]: value = "task-5113670" [ 683.626696] env[65522]: _type = "Task" [ 683.626696] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.647159] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113670, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.898282] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Acquiring lock "refresh_cache-68d8d59b-4fc8-4993-8c9e-31b2329c88dd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.898282] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Acquired lock "refresh_cache-68d8d59b-4fc8-4993-8c9e-31b2329c88dd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 683.898282] env[65522]: DEBUG nova.network.neutron [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 683.898282] env[65522]: DEBUG nova.scheduler.client.report [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 683.919748] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a58136-abfd-d1b0-7082-06cdf5143e18, 'name': SearchDatastore_Task, 'duration_secs': 0.010893} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.921029] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 683.921029] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 683.921261] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.921261] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 683.921436] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 683.925055] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ed5f878-b7d1-4f5a-8be4-e6efd0e2470f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.938893] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 683.939093] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 683.940017] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1dd655fe-63e3-4c7a-a137-15f9ea9e27ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.947559] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 683.947559] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52076af9-af4b-44ff-c3b4-dcd55a9c94bb" [ 683.947559] env[65522]: _type = "Task" [ 683.947559] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.958314] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52076af9-af4b-44ff-c3b4-dcd55a9c94bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.145590] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113670, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.179865] env[65522]: DEBUG nova.network.neutron [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Updating instance_info_cache with network_info: [{"id": "60d5d352-4fed-4b0f-9efc-ce13e2066444", "address": "fa:16:3e:72:1d:e5", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60d5d352-4f", "ovs_interfaceid": "60d5d352-4fed-4b0f-9efc-ce13e2066444", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 684.406355] env[65522]: WARNING openstack [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.406355] env[65522]: WARNING openstack [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.424200] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 684.431241] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.927s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.437353] env[65522]: INFO nova.compute.claims [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.475744] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52076af9-af4b-44ff-c3b4-dcd55a9c94bb, 'name': SearchDatastore_Task, 'duration_secs': 0.034717} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.477331] env[65522]: INFO nova.scheduler.client.report [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Deleted allocations for instance 3f749f68-879f-44f3-8477-7ecea947078c [ 684.484243] env[65522]: DEBUG nova.network.neutron [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 684.485450] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa76e74a-9557-4279-abaf-3ec095875895 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.499691] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 684.499691] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c4772c-2c5b-1266-a6f8-f2128b9713e9" [ 684.499691] env[65522]: _type = "Task" [ 684.499691] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.513313] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c4772c-2c5b-1266-a6f8-f2128b9713e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.536215] env[65522]: DEBUG nova.compute.manager [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Received event network-changed-7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 684.536447] env[65522]: DEBUG nova.compute.manager [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Refreshing instance network info cache due to event network-changed-7d49f270-6cbd-4767-91fc-f899c46564c9. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 684.536667] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Acquiring lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.536804] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Acquired lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 684.536954] env[65522]: DEBUG nova.network.neutron [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Refreshing network info cache for port 7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 684.555870] env[65522]: WARNING openstack [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.556284] env[65522]: WARNING openstack [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.646276] env[65522]: WARNING neutronclient.v2_0.client [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 684.646865] env[65522]: WARNING openstack [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.649605] env[65522]: WARNING openstack [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.662371] env[65522]: DEBUG oslo_vmware.api [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113670, 'name': PowerOnVM_Task, 'duration_secs': 0.862758} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.663522] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 684.663522] env[65522]: INFO nova.compute.manager [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Took 9.76 seconds to spawn the instance on the hypervisor. [ 684.663522] env[65522]: DEBUG nova.compute.manager [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 684.664012] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3694a83-dd90-4097-ae0a-e7d235a3fe52 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.682658] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "refresh_cache-87a3c63d-794d-44ab-bad6-65c323d72ae7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 684.684518] env[65522]: DEBUG nova.compute.manager [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Instance network_info: |[{"id": "60d5d352-4fed-4b0f-9efc-ce13e2066444", "address": "fa:16:3e:72:1d:e5", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60d5d352-4f", "ovs_interfaceid": "60d5d352-4fed-4b0f-9efc-ce13e2066444", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 684.684713] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:1d:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4b033f4d-2e92-4702-add6-410a29d3f251', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '60d5d352-4fed-4b0f-9efc-ce13e2066444', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 684.695895] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 684.699768] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 684.699768] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d09debfc-5bcc-4031-b87a-7fb847c1aae3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.723033] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 684.723033] env[65522]: value = "task-5113671" [ 684.723033] env[65522]: _type = "Task" [ 684.723033] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.733635] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113671, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.937021] env[65522]: DEBUG nova.network.neutron [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Updating instance_info_cache with network_info: [{"id": "3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6", "address": "fa:16:3e:6f:2f:ba", "network": {"id": "bed09e85-fdc4-4258-9be8-027aa4cf34a6", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-752246656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb9a9e5da1194ab7927f72d21d1484b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de5fcb06-b0d0-467f-86fe-06882165ac31", "external-id": "nsx-vlan-transportzone-595", "segmentation_id": 595, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bbf4b9c-9f", "ovs_interfaceid": "3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 685.007219] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91a8446e-3a4a-4ed9-9bec-5bf445c4018e tempest-TenantUsagesTestJSON-474643970 tempest-TenantUsagesTestJSON-474643970-project-member] Lock "3f749f68-879f-44f3-8477-7ecea947078c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.473s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 685.019734] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c4772c-2c5b-1266-a6f8-f2128b9713e9, 'name': SearchDatastore_Task, 'duration_secs': 0.013695} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.019734] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 685.021132] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] cd82f2fd-d657-4ece-8e90-9a47072f5546/cd82f2fd-d657-4ece-8e90-9a47072f5546.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 685.021132] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30b3b87a-5f81-4689-bab9-86c0b7790d67 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.033392] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 685.033392] env[65522]: value = "task-5113672" [ 685.033392] env[65522]: _type = "Task" [ 685.033392] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.040626] env[65522]: WARNING neutronclient.v2_0.client [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.041738] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.042145] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 685.053706] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113672, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.195072] env[65522]: INFO nova.compute.manager [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Took 23.87 seconds to build instance. [ 685.229916] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.230309] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 685.253639] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113671, 'name': CreateVM_Task, 'duration_secs': 0.502556} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.253819] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 685.254491] env[65522]: WARNING neutronclient.v2_0.client [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.254871] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.255049] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 685.255391] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 685.255685] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c33febae-5507-47b9-907a-c2ccc8255276 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.263244] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 685.263244] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529940ca-d387-e508-fe30-fd6e90ff9ea7" [ 685.263244] env[65522]: _type = "Task" [ 685.263244] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.272954] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529940ca-d387-e508-fe30-fd6e90ff9ea7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.409925] env[65522]: WARNING neutronclient.v2_0.client [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.410887] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.410967] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 685.440234] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Releasing lock "refresh_cache-68d8d59b-4fc8-4993-8c9e-31b2329c88dd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 685.440513] env[65522]: DEBUG nova.compute.manager [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Instance network_info: |[{"id": "3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6", "address": "fa:16:3e:6f:2f:ba", "network": {"id": "bed09e85-fdc4-4258-9be8-027aa4cf34a6", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-752246656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb9a9e5da1194ab7927f72d21d1484b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de5fcb06-b0d0-467f-86fe-06882165ac31", "external-id": "nsx-vlan-transportzone-595", "segmentation_id": 595, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bbf4b9c-9f", "ovs_interfaceid": "3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 685.441062] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:2f:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'de5fcb06-b0d0-467f-86fe-06882165ac31', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 685.455947] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Creating folder: Project (eb9a9e5da1194ab7927f72d21d1484b0). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 685.456757] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ccfe60c-1f89-4bc6-8937-8c9db25c6568 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.471555] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Created folder: Project (eb9a9e5da1194ab7927f72d21d1484b0) in parent group-v994660. [ 685.471750] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Creating folder: Instances. Parent ref: group-v994722. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 685.472036] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-08589ed0-89c0-4790-82b5-5bf3b031d3a9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.489583] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Created folder: Instances in parent group-v994722. [ 685.489583] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 685.489583] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 685.489583] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-455e24e0-5e17-47a9-bd99-407a54c06fe2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.522520] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 685.522520] env[65522]: value = "task-5113675" [ 685.522520] env[65522]: _type = "Task" [ 685.522520] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.535796] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113675, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.548822] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113672, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.698767] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a0d7f454-1b5a-464c-8ad1-134c7d86259b tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "ca964440-5375-4aff-8b45-96fbe829dd16" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.390s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 685.778582] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529940ca-d387-e508-fe30-fd6e90ff9ea7, 'name': SearchDatastore_Task, 'duration_secs': 0.013489} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.783084] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 685.783490] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 685.783896] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.784069] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 685.784451] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 685.785279] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd47e13b-5f9e-4e98-bfad-fcadab29411c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.806381] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 685.806693] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 685.815248] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61b9de06-1aef-4088-ac26-920c16c70094 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.825052] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 685.825052] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f71e83-c851-2af8-2bf9-418db5e13bef" [ 685.825052] env[65522]: _type = "Task" [ 685.825052] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.844338] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f71e83-c851-2af8-2bf9-418db5e13bef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.855025] env[65522]: DEBUG nova.network.neutron [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Updated VIF entry in instance network info cache for port 7d49f270-6cbd-4767-91fc-f899c46564c9. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 685.855396] env[65522]: DEBUG nova.network.neutron [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Updating instance_info_cache with network_info: [{"id": "7d49f270-6cbd-4767-91fc-f899c46564c9", "address": "fa:16:3e:af:04:9f", "network": {"id": "6a60138b-7f03-4817-ac9a-1facdc1cc0cc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1344754077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4fdbd0ed78c41a591ccc0a587313ec8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f85835c8-5d0c-4b2f-97c4-6c4006580f79", "external-id": "nsx-vlan-transportzone-245", "segmentation_id": 245, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d49f270-6c", "ovs_interfaceid": "7d49f270-6cbd-4767-91fc-f899c46564c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 685.975982] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073974b5-f298-4bde-b20c-e3e463fb8d67 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.988372] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed48bb4c-38e5-4075-8b98-41ed936e0598 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.783322] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Releasing lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 686.783633] env[65522]: DEBUG nova.compute.manager [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Received event network-changed-aca7a84f-d08f-4c79-821e-1f90b00f141b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 686.784050] env[65522]: DEBUG nova.compute.manager [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Refreshing instance network info cache due to event network-changed-aca7a84f-d08f-4c79-821e-1f90b00f141b. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 686.784050] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Acquiring lock "refresh_cache-ad074365-1443-481d-990f-479c523219fa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.784050] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Acquired lock "refresh_cache-ad074365-1443-481d-990f-479c523219fa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 686.784215] env[65522]: DEBUG nova.network.neutron [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Refreshing network info cache for port aca7a84f-d08f-4c79-821e-1f90b00f141b {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 686.797829] env[65522]: DEBUG nova.compute.manager [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Received event network-changed-ede8edfa-842b-482d-8e7b-34bde13a02f8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 686.798043] env[65522]: DEBUG nova.compute.manager [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Refreshing instance network info cache due to event network-changed-ede8edfa-842b-482d-8e7b-34bde13a02f8. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 686.798298] env[65522]: DEBUG oslo_concurrency.lockutils [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Acquiring lock "refresh_cache-ca964440-5375-4aff-8b45-96fbe829dd16" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.798421] env[65522]: DEBUG oslo_concurrency.lockutils [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Acquired lock "refresh_cache-ca964440-5375-4aff-8b45-96fbe829dd16" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 686.798569] env[65522]: DEBUG nova.network.neutron [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Refreshing network info cache for port ede8edfa-842b-482d-8e7b-34bde13a02f8 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 686.800457] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b045cc-2df6-4b59-a2a0-a1affa21bc82 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.818514] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f71e83-c851-2af8-2bf9-418db5e13bef, 'name': SearchDatastore_Task, 'duration_secs': 0.01499} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.818758] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113672, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.804563} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.818909] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113675, 'name': CreateVM_Task, 'duration_secs': 0.80299} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.822288] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] cd82f2fd-d657-4ece-8e90-9a47072f5546/cd82f2fd-d657-4ece-8e90-9a47072f5546.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 686.822288] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 686.822288] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 686.822495] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f123a97-ec02-4c07-9553-f1ae264669e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.825795] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-04ddccc6-77ce-49b9-bf4e-be12fdf639c9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.828273] env[65522]: WARNING neutronclient.v2_0.client [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 686.829547] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.829547] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 686.829547] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 686.830300] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bac0bcb-cdc7-4155-b250-31fcee4a6f38 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.834883] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92c77add-944c-4576-b8e7-1e170fda8b16 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.841485] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 686.841485] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5292b653-424d-9dd5-a37b-eadc6441ea1c" [ 686.841485] env[65522]: _type = "Task" [ 686.841485] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.852841] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Waiting for the task: (returnval){ [ 686.852841] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f7515f-b7d9-3716-6586-1f1712f89c69" [ 686.852841] env[65522]: _type = "Task" [ 686.852841] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.853464] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 686.853464] env[65522]: value = "task-5113680" [ 686.853464] env[65522]: _type = "Task" [ 686.853464] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.854203] env[65522]: DEBUG nova.compute.provider_tree [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.873691] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f7515f-b7d9-3716-6586-1f1712f89c69, 'name': SearchDatastore_Task, 'duration_secs': 0.021464} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.881426] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 686.881562] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 686.881803] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.882277] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113680, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.882842] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5292b653-424d-9dd5-a37b-eadc6441ea1c, 'name': SearchDatastore_Task, 'duration_secs': 0.022878} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.882842] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 686.882975] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 87a3c63d-794d-44ab-bad6-65c323d72ae7/87a3c63d-794d-44ab-bad6-65c323d72ae7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 686.883205] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 686.883369] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 686.883573] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5612362f-7afe-4284-a9bc-1eaf4652cf2d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.886713] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6cc4a2e4-a6ce-4a6e-b0ed-9c319baedc29 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.893620] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "f44db343-7e17-4780-9a67-245992ac1a63" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 686.893928] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "f44db343-7e17-4780-9a67-245992ac1a63" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 686.899993] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 686.899993] env[65522]: value = "task-5113681" [ 686.899993] env[65522]: _type = "Task" [ 686.899993] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.907261] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 686.907261] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 686.908655] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beb65974-5d9b-4d55-94db-413ea3f88766 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.915683] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113681, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.920344] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Waiting for the task: (returnval){ [ 686.920344] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52287f57-0e68-c487-755f-c3a69212ba79" [ 686.920344] env[65522]: _type = "Task" [ 686.920344] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.932315] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52287f57-0e68-c487-755f-c3a69212ba79, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.288409] env[65522]: WARNING neutronclient.v2_0.client [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 687.289240] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.289635] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.308867] env[65522]: WARNING neutronclient.v2_0.client [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 687.309750] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.310133] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.364017] env[65522]: DEBUG nova.scheduler.client.report [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 687.378925] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113680, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.15678} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.379266] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 687.380237] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81da7d1-abc5-4ab7-b378-8e8715899af2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.398350] env[65522]: DEBUG nova.compute.manager [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 687.414034] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Reconfiguring VM instance instance-00000015 to attach disk [datastore2] cd82f2fd-d657-4ece-8e90-9a47072f5546/cd82f2fd-d657-4ece-8e90-9a47072f5546.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 687.414034] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac865388-6e07-46b7-b7a7-a71beab02ce8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.456504] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52287f57-0e68-c487-755f-c3a69212ba79, 'name': SearchDatastore_Task, 'duration_secs': 0.023779} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.462724] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113681, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.464466] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 687.464466] env[65522]: value = "task-5113682" [ 687.464466] env[65522]: _type = "Task" [ 687.464466] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.464466] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f2712b7-e9f6-4025-9911-b44a36c87e14 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.476317] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Waiting for the task: (returnval){ [ 687.476317] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529e9d6f-dd90-0181-9220-0fab43f0c6df" [ 687.476317] env[65522]: _type = "Task" [ 687.476317] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.481575] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113682, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.497134] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529e9d6f-dd90-0181-9220-0fab43f0c6df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.871761] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.440s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 687.872324] env[65522]: DEBUG nova.compute.manager [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 687.875775] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.928s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 687.878175] env[65522]: INFO nova.compute.claims [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.937293] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113681, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.641036} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.938556] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 87a3c63d-794d-44ab-bad6-65c323d72ae7/87a3c63d-794d-44ab-bad6-65c323d72ae7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 687.939937] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 687.939937] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84105304-af13-44ea-a119-c05422bd2ccd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.951289] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 687.951897] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 687.951897] env[65522]: value = "task-5113683" [ 687.951897] env[65522]: _type = "Task" [ 687.951897] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.966052] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113683, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.976226] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113682, 'name': ReconfigVM_Task, 'duration_secs': 0.45547} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.976566] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Reconfigured VM instance instance-00000015 to attach disk [datastore2] cd82f2fd-d657-4ece-8e90-9a47072f5546/cd82f2fd-d657-4ece-8e90-9a47072f5546.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 687.977207] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0c4e6b3-9970-484c-80c1-c06e706824da {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.986337] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 687.986337] env[65522]: value = "task-5113684" [ 687.986337] env[65522]: _type = "Task" [ 687.986337] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.993088] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529e9d6f-dd90-0181-9220-0fab43f0c6df, 'name': SearchDatastore_Task, 'duration_secs': 0.04904} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.993783] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 687.994087] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 68d8d59b-4fc8-4993-8c9e-31b2329c88dd/68d8d59b-4fc8-4993-8c9e-31b2329c88dd.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 687.994384] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-833884e7-1fd6-42ef-a877-790688146e63 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.001159] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113684, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.012641] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Waiting for the task: (returnval){ [ 688.012641] env[65522]: value = "task-5113685" [ 688.012641] env[65522]: _type = "Task" [ 688.012641] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.023604] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5113685, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.385559] env[65522]: DEBUG nova.compute.utils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 688.397878] env[65522]: DEBUG nova.compute.manager [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 688.397878] env[65522]: DEBUG nova.network.neutron [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 688.397878] env[65522]: WARNING neutronclient.v2_0.client [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 688.397878] env[65522]: WARNING neutronclient.v2_0.client [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 688.397878] env[65522]: WARNING openstack [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.398337] env[65522]: WARNING openstack [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.455038] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.455222] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.484244] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113683, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.199867} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.484244] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 688.485545] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e657fb2d-164d-4eb8-861f-b77f3a8512c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.493263] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.493973] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.513631] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113684, 'name': Rename_Task, 'duration_secs': 0.324509} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.526988] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 688.535784] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Reconfiguring VM instance instance-00000016 to attach disk [datastore2] 87a3c63d-794d-44ab-bad6-65c323d72ae7/87a3c63d-794d-44ab-bad6-65c323d72ae7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 688.536279] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe526eaf-a18d-4626-b394-c2798783e258 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.538038] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3e6d72a-fcdd-4b4e-8b9b-f265016b179c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.562831] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5113685, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.566347] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 688.566347] env[65522]: value = "task-5113686" [ 688.566347] env[65522]: _type = "Task" [ 688.566347] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.566671] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 688.566671] env[65522]: value = "task-5113687" [ 688.566671] env[65522]: _type = "Task" [ 688.566671] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.579297] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113686, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.582737] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113687, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.839234] env[65522]: DEBUG nova.policy [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e74c132f591d4c6fbe65a75c1e1e4a24', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ecccb656b0d4c96b40b200cdcddbad5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 688.896029] env[65522]: DEBUG nova.compute.manager [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 689.027176] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5113685, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.990305} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.027491] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 68d8d59b-4fc8-4993-8c9e-31b2329c88dd/68d8d59b-4fc8-4993-8c9e-31b2329c88dd.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 689.027744] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 689.027940] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dfb882bc-978d-4bf1-b803-84790f43fe21 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.035886] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Waiting for the task: (returnval){ [ 689.035886] env[65522]: value = "task-5113688" [ 689.035886] env[65522]: _type = "Task" [ 689.035886] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.053029] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5113688, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.086385] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113686, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.090032] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113687, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.353406] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9120db9e-4ab1-4914-93cd-1f46a14971a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.363511] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937071c0-7846-4df3-a525-a89d7ca4a786 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.401502] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89d364b-94ca-405a-987c-c94f9a2bdcaa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.413478] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21f12cd-fb5d-4bf4-9eae-7df0a68b8843 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.429235] env[65522]: DEBUG nova.compute.provider_tree [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.547783] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5113688, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.120007} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.547783] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 689.548446] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd888ea9-c131-47d2-8030-1f80d3e50544 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.573080] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Reconfiguring VM instance instance-00000017 to attach disk [datastore2] 68d8d59b-4fc8-4993-8c9e-31b2329c88dd/68d8d59b-4fc8-4993-8c9e-31b2329c88dd.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 689.573902] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d6106f0-7fc6-4a62-a579-389348b2cb2f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.600535] env[65522]: DEBUG oslo_vmware.api [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113686, 'name': PowerOnVM_Task, 'duration_secs': 0.719088} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.604367] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 689.605228] env[65522]: INFO nova.compute.manager [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Took 12.24 seconds to spawn the instance on the hypervisor. [ 689.605228] env[65522]: DEBUG nova.compute.manager [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 689.605228] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Waiting for the task: (returnval){ [ 689.605228] env[65522]: value = "task-5113690" [ 689.605228] env[65522]: _type = "Task" [ 689.605228] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.605474] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113687, 'name': ReconfigVM_Task, 'duration_secs': 0.546783} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.606102] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544e7019-462a-4e2c-819c-dfa86af728bf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.609621] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Reconfigured VM instance instance-00000016 to attach disk [datastore2] 87a3c63d-794d-44ab-bad6-65c323d72ae7/87a3c63d-794d-44ab-bad6-65c323d72ae7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 689.612122] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87613a17-e453-4c0e-aaa5-eb4a5c4a30fc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.622113] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5113690, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.627094] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 689.627094] env[65522]: value = "task-5113691" [ 689.627094] env[65522]: _type = "Task" [ 689.627094] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.635738] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113691, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.850617] env[65522]: DEBUG nova.network.neutron [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Successfully created port: 189bde05-0b1d-4c6a-be88-2453345dd731 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 689.902146] env[65522]: WARNING neutronclient.v2_0.client [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 689.902844] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.903214] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.913209] env[65522]: DEBUG nova.compute.manager [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 689.920082] env[65522]: WARNING neutronclient.v2_0.client [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 689.920823] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.921388] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.932801] env[65522]: DEBUG nova.scheduler.client.report [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 689.944179] env[65522]: DEBUG nova.virt.hardware [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 689.944468] env[65522]: DEBUG nova.virt.hardware [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 689.944646] env[65522]: DEBUG nova.virt.hardware [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 689.944771] env[65522]: DEBUG nova.virt.hardware [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 689.944912] env[65522]: DEBUG nova.virt.hardware [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 689.945069] env[65522]: DEBUG nova.virt.hardware [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 689.945293] env[65522]: DEBUG nova.virt.hardware [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 689.945460] env[65522]: DEBUG nova.virt.hardware [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 689.945610] env[65522]: DEBUG nova.virt.hardware [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 689.945767] env[65522]: DEBUG nova.virt.hardware [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 689.945931] env[65522]: DEBUG nova.virt.hardware [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 689.947025] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc19b225-8d74-45c4-bf5d-fe377b8fc7ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.957746] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d74d692-c6e2-40f5-9e9f-0757011ef50f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.122582] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5113690, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.141881] env[65522]: INFO nova.compute.manager [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Took 27.94 seconds to build instance. [ 690.149423] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113691, 'name': Rename_Task, 'duration_secs': 0.169844} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.149798] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 690.150079] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f097f75-137e-4911-acef-f753a4ff2210 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.163042] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 690.163042] env[65522]: value = "task-5113692" [ 690.163042] env[65522]: _type = "Task" [ 690.163042] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.179457] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113692, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.438861] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 690.439069] env[65522]: DEBUG nova.compute.manager [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 690.442539] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.476s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 690.442539] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 690.442539] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 690.442539] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.414s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 690.444154] env[65522]: INFO nova.compute.claims [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.447952] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728f9f81-fcd7-43c2-b7d4-b3e09297a57a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.458738] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d42a5ef-1ae1-440b-836b-c7465b18aa15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.475992] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a7afff4-bec4-4b40-bbd2-29b772d4006f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.485140] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b966859f-067c-4b40-a385-d5cb1731d425 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.523021] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180231MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 690.523021] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 690.620620] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5113690, 'name': ReconfigVM_Task, 'duration_secs': 0.83773} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.620882] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Reconfigured VM instance instance-00000017 to attach disk [datastore2] 68d8d59b-4fc8-4993-8c9e-31b2329c88dd/68d8d59b-4fc8-4993-8c9e-31b2329c88dd.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 690.621624] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-11aff973-9088-4972-92d4-c6f293beeebc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.631341] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Waiting for the task: (returnval){ [ 690.631341] env[65522]: value = "task-5113693" [ 690.631341] env[65522]: _type = "Task" [ 690.631341] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.644398] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55aca080-40f7-4e56-8c3d-bef012d03ef3 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "cd82f2fd-d657-4ece-8e90-9a47072f5546" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.462s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 690.644621] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5113693, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.675277] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113692, 'name': PowerOnVM_Task} progress is 90%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.949505] env[65522]: DEBUG nova.compute.utils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 690.960333] env[65522]: DEBUG nova.compute.manager [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 690.960532] env[65522]: DEBUG nova.network.neutron [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 690.961151] env[65522]: WARNING neutronclient.v2_0.client [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.961151] env[65522]: WARNING neutronclient.v2_0.client [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.961667] env[65522]: WARNING openstack [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.962115] env[65522]: WARNING openstack [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.972300] env[65522]: DEBUG nova.compute.manager [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 691.144511] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5113693, 'name': Rename_Task, 'duration_secs': 0.336068} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.144817] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 691.145098] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0789feb5-fb31-4466-a3c5-00f8001e4ad1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.153137] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Waiting for the task: (returnval){ [ 691.153137] env[65522]: value = "task-5113694" [ 691.153137] env[65522]: _type = "Task" [ 691.153137] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.166507] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5113694, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.176600] env[65522]: DEBUG oslo_vmware.api [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113692, 'name': PowerOnVM_Task, 'duration_secs': 0.569488} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.176906] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 691.177119] env[65522]: INFO nova.compute.manager [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Took 11.24 seconds to spawn the instance on the hypervisor. [ 691.177333] env[65522]: DEBUG nova.compute.manager [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 691.178159] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c2791a-5718-44bf-ae71-3d31f0a80c58 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.667923] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5113694, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.707682] env[65522]: INFO nova.compute.manager [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Took 25.27 seconds to build instance. [ 691.893151] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1de009-0a75-415f-9aab-37308bacaa95 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.907222] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9486ff7-57c8-4436-9d17-b8f6d085a87c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.944647] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9179ef5-041b-4764-b085-a8065e1882d2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.954180] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81117ef5-647d-40af-91e5-1d0ab3bd61ca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.973156] env[65522]: DEBUG nova.compute.provider_tree [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 691.985465] env[65522]: DEBUG nova.compute.manager [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 692.025150] env[65522]: DEBUG nova.virt.hardware [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 692.025783] env[65522]: DEBUG nova.virt.hardware [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 692.026061] env[65522]: DEBUG nova.virt.hardware [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 692.026302] env[65522]: DEBUG nova.virt.hardware [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 692.026459] env[65522]: DEBUG nova.virt.hardware [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 692.026617] env[65522]: DEBUG nova.virt.hardware [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 692.027222] env[65522]: DEBUG nova.virt.hardware [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 692.027597] env[65522]: DEBUG nova.virt.hardware [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 692.027673] env[65522]: DEBUG nova.virt.hardware [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 692.027838] env[65522]: DEBUG nova.virt.hardware [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 692.028095] env[65522]: DEBUG nova.virt.hardware [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 692.029163] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a15397f9-10df-41f3-bbbd-0eae76974e59 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.043890] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08fd8ad2-4fe8-4500-9345-af5165534468 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.166904] env[65522]: DEBUG oslo_vmware.api [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5113694, 'name': PowerOnVM_Task, 'duration_secs': 0.896351} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.167210] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 692.167413] env[65522]: INFO nova.compute.manager [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Took 9.73 seconds to spawn the instance on the hypervisor. [ 692.168197] env[65522]: DEBUG nova.compute.manager [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 692.168563] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20028a3a-7de6-44ff-b125-9c4a9d7ecf6f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.213299] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b43c933-dd31-42cf-aaf9-27b434c93a04 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "87a3c63d-794d-44ab-bad6-65c323d72ae7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.792s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 692.348966] env[65522]: DEBUG nova.network.neutron [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Successfully updated port: 189bde05-0b1d-4c6a-be88-2453345dd731 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 692.446035] env[65522]: DEBUG nova.network.neutron [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Updated VIF entry in instance network info cache for port aca7a84f-d08f-4c79-821e-1f90b00f141b. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 692.446437] env[65522]: DEBUG nova.network.neutron [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Updating instance_info_cache with network_info: [{"id": "aca7a84f-d08f-4c79-821e-1f90b00f141b", "address": "fa:16:3e:99:c9:27", "network": {"id": "6a60138b-7f03-4817-ac9a-1facdc1cc0cc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1344754077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4fdbd0ed78c41a591ccc0a587313ec8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f85835c8-5d0c-4b2f-97c4-6c4006580f79", "external-id": "nsx-vlan-transportzone-245", "segmentation_id": 245, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaca7a84f-d0", "ovs_interfaceid": "aca7a84f-d08f-4c79-821e-1f90b00f141b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 692.488023] env[65522]: DEBUG nova.network.neutron [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Updated VIF entry in instance network info cache for port ede8edfa-842b-482d-8e7b-34bde13a02f8. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 692.488023] env[65522]: DEBUG nova.network.neutron [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Updating instance_info_cache with network_info: [{"id": "ede8edfa-842b-482d-8e7b-34bde13a02f8", "address": "fa:16:3e:f6:50:03", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapede8edfa-84", "ovs_interfaceid": "ede8edfa-842b-482d-8e7b-34bde13a02f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 692.507761] env[65522]: ERROR nova.scheduler.client.report [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [req-bb8934ce-a4c5-4f9c-bbcc-1caa15445b7f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9a0d1ac8-ed26-4657-8899-fbceab29db0d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-bb8934ce-a4c5-4f9c-bbcc-1caa15445b7f"}]} [ 692.538766] env[65522]: DEBUG nova.scheduler.client.report [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Refreshing inventories for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 692.562338] env[65522]: DEBUG nova.scheduler.client.report [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Updating ProviderTree inventory for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 692.563033] env[65522]: DEBUG nova.compute.provider_tree [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 692.578835] env[65522]: DEBUG nova.scheduler.client.report [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Refreshing aggregate associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, aggregates: None {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 692.611851] env[65522]: DEBUG nova.scheduler.client.report [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Refreshing trait associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 692.673073] env[65522]: DEBUG nova.policy [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1103ba172d754007bb2b7339e1d674f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '503adb79a1f74f1b98e1f06ec54f1584', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 692.704138] env[65522]: INFO nova.compute.manager [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Took 24.09 seconds to build instance. [ 692.859040] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "refresh_cache-f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.859040] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "refresh_cache-f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 692.859040] env[65522]: DEBUG nova.network.neutron [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 692.952377] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Releasing lock "refresh_cache-ad074365-1443-481d-990f-479c523219fa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 692.952645] env[65522]: DEBUG nova.compute.manager [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Received event network-vif-plugged-ff1e8448-0d72-4738-ae2b-9b543dbe9f5e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 692.952839] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Acquiring lock "cd82f2fd-d657-4ece-8e90-9a47072f5546-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 692.953043] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Lock "cd82f2fd-d657-4ece-8e90-9a47072f5546-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 692.953237] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Lock "cd82f2fd-d657-4ece-8e90-9a47072f5546-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 692.953361] env[65522]: DEBUG nova.compute.manager [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] No waiting events found dispatching network-vif-plugged-ff1e8448-0d72-4738-ae2b-9b543dbe9f5e {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 692.953629] env[65522]: WARNING nova.compute.manager [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Received unexpected event network-vif-plugged-ff1e8448-0d72-4738-ae2b-9b543dbe9f5e for instance with vm_state building and task_state spawning. [ 692.953710] env[65522]: DEBUG nova.compute.manager [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Received event network-changed-ff1e8448-0d72-4738-ae2b-9b543dbe9f5e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 692.953816] env[65522]: DEBUG nova.compute.manager [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Refreshing instance network info cache due to event network-changed-ff1e8448-0d72-4738-ae2b-9b543dbe9f5e. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 692.953989] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Acquiring lock "refresh_cache-cd82f2fd-d657-4ece-8e90-9a47072f5546" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.954141] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Acquired lock "refresh_cache-cd82f2fd-d657-4ece-8e90-9a47072f5546" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 692.954430] env[65522]: DEBUG nova.network.neutron [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Refreshing network info cache for port ff1e8448-0d72-4738-ae2b-9b543dbe9f5e {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 692.992402] env[65522]: DEBUG oslo_concurrency.lockutils [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Releasing lock "refresh_cache-ca964440-5375-4aff-8b45-96fbe829dd16" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 692.992402] env[65522]: DEBUG nova.compute.manager [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Received event network-changed-a7a9fee0-67af-4446-b838-0cd8922465b9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 692.992402] env[65522]: DEBUG nova.compute.manager [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Refreshing instance network info cache due to event network-changed-a7a9fee0-67af-4446-b838-0cd8922465b9. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 692.992762] env[65522]: DEBUG oslo_concurrency.lockutils [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Acquiring lock "refresh_cache-fdf56fa6-ad18-499e-a171-6242abb293bf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.992762] env[65522]: DEBUG oslo_concurrency.lockutils [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Acquired lock "refresh_cache-fdf56fa6-ad18-499e-a171-6242abb293bf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 692.993298] env[65522]: DEBUG nova.network.neutron [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Refreshing network info cache for port a7a9fee0-67af-4446-b838-0cd8922465b9 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 693.069551] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04018f3e-e254-49a5-a159-4ed8e6810980 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.080468] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f2b34a2-0921-4077-a832-4f37237bc869 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.113403] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1630ab21-edcf-4440-ac61-8464586ef9aa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.123184] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6d174b-2eda-4dde-86b8-1cef3f6e9c78 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.142270] env[65522]: DEBUG nova.compute.provider_tree [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 693.205925] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2058e3f5-e2e0-4bf3-b03a-4b84dc650f05 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Lock "68d8d59b-4fc8-4993-8c9e-31b2329c88dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.602s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 693.276083] env[65522]: DEBUG nova.network.neutron [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Successfully created port: fb1cd05a-757b-469c-9749-fd1ca4fcb168 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 693.354311] env[65522]: DEBUG nova.compute.manager [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Received event network-changed-97dc7703-dd4a-43dd-8e2d-02f8a31d1a69 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 693.354311] env[65522]: DEBUG nova.compute.manager [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Refreshing instance network info cache due to event network-changed-97dc7703-dd4a-43dd-8e2d-02f8a31d1a69. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 693.354311] env[65522]: DEBUG oslo_concurrency.lockutils [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] Acquiring lock "refresh_cache-0d52b207-ac69-48ce-b3af-7a83a499886e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.354311] env[65522]: DEBUG oslo_concurrency.lockutils [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] Acquired lock "refresh_cache-0d52b207-ac69-48ce-b3af-7a83a499886e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 693.354311] env[65522]: DEBUG nova.network.neutron [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Refreshing network info cache for port 97dc7703-dd4a-43dd-8e2d-02f8a31d1a69 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 693.358878] env[65522]: WARNING openstack [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.359565] env[65522]: WARNING openstack [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.406828] env[65522]: DEBUG nova.network.neutron [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 693.457682] env[65522]: WARNING neutronclient.v2_0.client [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 693.458417] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.459116] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.500065] env[65522]: WARNING openstack [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.500482] env[65522]: WARNING openstack [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.508221] env[65522]: WARNING neutronclient.v2_0.client [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 693.509023] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.509469] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.718800] env[65522]: DEBUG nova.scheduler.client.report [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Updated inventory for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with generation 32 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 693.718800] env[65522]: DEBUG nova.compute.provider_tree [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Updating resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d generation from 32 to 33 during operation: update_inventory {{(pid=65522) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 693.718800] env[65522]: DEBUG nova.compute.provider_tree [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 693.801549] env[65522]: WARNING neutronclient.v2_0.client [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 693.802504] env[65522]: WARNING openstack [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.802882] env[65522]: WARNING openstack [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.856394] env[65522]: WARNING neutronclient.v2_0.client [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 693.858520] env[65522]: WARNING openstack [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.860711] env[65522]: WARNING openstack [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.928313] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.928710] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.047344] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.047778] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.190016] env[65522]: DEBUG nova.network.neutron [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Updating instance_info_cache with network_info: [{"id": "189bde05-0b1d-4c6a-be88-2453345dd731", "address": "fa:16:3e:76:53:d6", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap189bde05-0b", "ovs_interfaceid": "189bde05-0b1d-4c6a-be88-2453345dd731", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 694.226558] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.783s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 694.226558] env[65522]: DEBUG nova.compute.manager [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 694.238338] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.287s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 694.240905] env[65522]: INFO nova.compute.claims [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 694.547524] env[65522]: WARNING neutronclient.v2_0.client [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 694.547524] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.548194] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.592247] env[65522]: WARNING openstack [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.592636] env[65522]: WARNING openstack [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.619765] env[65522]: WARNING neutronclient.v2_0.client [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 694.620398] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.620763] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.693489] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "refresh_cache-f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 694.693887] env[65522]: DEBUG nova.compute.manager [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Instance network_info: |[{"id": "189bde05-0b1d-4c6a-be88-2453345dd731", "address": "fa:16:3e:76:53:d6", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap189bde05-0b", "ovs_interfaceid": "189bde05-0b1d-4c6a-be88-2453345dd731", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 694.694551] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:53:d6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6fb0104-186b-4288-b87e-634893f46f01', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '189bde05-0b1d-4c6a-be88-2453345dd731', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 694.708476] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 694.708880] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 694.709521] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11236fec-1b43-4c7d-9a2f-17f3f15b203b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.742301] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 694.742301] env[65522]: value = "task-5113697" [ 694.742301] env[65522]: _type = "Task" [ 694.742301] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.747684] env[65522]: DEBUG nova.compute.utils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 694.752168] env[65522]: DEBUG nova.compute.manager [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 694.752168] env[65522]: DEBUG nova.network.neutron [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 694.752512] env[65522]: WARNING neutronclient.v2_0.client [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 694.752873] env[65522]: WARNING neutronclient.v2_0.client [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 694.753618] env[65522]: WARNING openstack [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.753985] env[65522]: WARNING openstack [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.770390] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113697, 'name': CreateVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.021587] env[65522]: DEBUG nova.network.neutron [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Successfully updated port: fb1cd05a-757b-469c-9749-fd1ca4fcb168 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 695.123507] env[65522]: DEBUG nova.network.neutron [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Updated VIF entry in instance network info cache for port ff1e8448-0d72-4738-ae2b-9b543dbe9f5e. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 695.124063] env[65522]: DEBUG nova.network.neutron [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Updating instance_info_cache with network_info: [{"id": "ff1e8448-0d72-4738-ae2b-9b543dbe9f5e", "address": "fa:16:3e:b2:40:68", "network": {"id": "6314277f-6ed5-45e5-950d-2683782849aa", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-246048775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95bca233019f4e2f8bac0dd800392a69", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff1e8448-0d", "ovs_interfaceid": "ff1e8448-0d72-4738-ae2b-9b543dbe9f5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 695.126482] env[65522]: DEBUG oslo_concurrency.lockutils [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquiring lock "6e10e846-cdb7-458e-b511-18ab1742228e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 695.126817] env[65522]: DEBUG oslo_concurrency.lockutils [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Lock "6e10e846-cdb7-458e-b511-18ab1742228e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 695.180218] env[65522]: WARNING neutronclient.v2_0.client [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 695.180988] env[65522]: WARNING openstack [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 695.181351] env[65522]: WARNING openstack [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 695.220238] env[65522]: DEBUG nova.network.neutron [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Updated VIF entry in instance network info cache for port a7a9fee0-67af-4446-b838-0cd8922465b9. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 695.221447] env[65522]: DEBUG nova.network.neutron [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Updating instance_info_cache with network_info: [{"id": "a7a9fee0-67af-4446-b838-0cd8922465b9", "address": "fa:16:3e:33:16:af", "network": {"id": "29f88cc3-e21a-4a6e-87a2-8624a1be6482", "bridge": "br-int", "label": "tempest-ServersTestJSON-1088062070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5f64db4fc2f3412986ef6fd1393e99c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7a9fee0-67", "ovs_interfaceid": "a7a9fee0-67af-4446-b838-0cd8922465b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 695.228548] env[65522]: DEBUG nova.policy [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6dc2c46256514d9c87709cea348d64c5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '68c8ff7c2b604c939a2124142d1654b8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 695.256721] env[65522]: DEBUG nova.compute.manager [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 695.273465] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113697, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.521826] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.521998] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 695.522278] env[65522]: DEBUG nova.network.neutron [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 695.629827] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Releasing lock "refresh_cache-cd82f2fd-d657-4ece-8e90-9a47072f5546" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 695.630088] env[65522]: DEBUG nova.compute.manager [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Received event network-vif-plugged-60d5d352-4fed-4b0f-9efc-ce13e2066444 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 695.630309] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Acquiring lock "87a3c63d-794d-44ab-bad6-65c323d72ae7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 695.630549] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Lock "87a3c63d-794d-44ab-bad6-65c323d72ae7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 695.630763] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Lock "87a3c63d-794d-44ab-bad6-65c323d72ae7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 695.630922] env[65522]: DEBUG nova.compute.manager [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] No waiting events found dispatching network-vif-plugged-60d5d352-4fed-4b0f-9efc-ce13e2066444 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 695.631085] env[65522]: WARNING nova.compute.manager [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Received unexpected event network-vif-plugged-60d5d352-4fed-4b0f-9efc-ce13e2066444 for instance with vm_state building and task_state spawning. [ 695.631323] env[65522]: DEBUG nova.compute.manager [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Received event network-changed-60d5d352-4fed-4b0f-9efc-ce13e2066444 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 695.631612] env[65522]: DEBUG nova.compute.manager [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Refreshing instance network info cache due to event network-changed-60d5d352-4fed-4b0f-9efc-ce13e2066444. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 695.631712] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Acquiring lock "refresh_cache-87a3c63d-794d-44ab-bad6-65c323d72ae7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.631833] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Acquired lock "refresh_cache-87a3c63d-794d-44ab-bad6-65c323d72ae7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 695.631986] env[65522]: DEBUG nova.network.neutron [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Refreshing network info cache for port 60d5d352-4fed-4b0f-9efc-ce13e2066444 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 695.636129] env[65522]: DEBUG nova.compute.manager [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 695.723680] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e97f334-c642-4bf9-8dee-bd6db950aebd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.735288] env[65522]: DEBUG oslo_concurrency.lockutils [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Releasing lock "refresh_cache-fdf56fa6-ad18-499e-a171-6242abb293bf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 695.735561] env[65522]: DEBUG nova.compute.manager [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Received event network-vif-plugged-3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 695.735860] env[65522]: DEBUG oslo_concurrency.lockutils [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Acquiring lock "68d8d59b-4fc8-4993-8c9e-31b2329c88dd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 695.735945] env[65522]: DEBUG oslo_concurrency.lockutils [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Lock "68d8d59b-4fc8-4993-8c9e-31b2329c88dd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 695.736107] env[65522]: DEBUG oslo_concurrency.lockutils [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Lock "68d8d59b-4fc8-4993-8c9e-31b2329c88dd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 695.736292] env[65522]: DEBUG nova.compute.manager [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] No waiting events found dispatching network-vif-plugged-3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 695.736459] env[65522]: WARNING nova.compute.manager [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Received unexpected event network-vif-plugged-3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6 for instance with vm_state building and task_state spawning. [ 695.736623] env[65522]: DEBUG nova.compute.manager [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Received event network-changed-3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 695.736773] env[65522]: DEBUG nova.compute.manager [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Refreshing instance network info cache due to event network-changed-3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 695.736963] env[65522]: DEBUG oslo_concurrency.lockutils [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Acquiring lock "refresh_cache-68d8d59b-4fc8-4993-8c9e-31b2329c88dd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.737187] env[65522]: DEBUG oslo_concurrency.lockutils [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Acquired lock "refresh_cache-68d8d59b-4fc8-4993-8c9e-31b2329c88dd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 695.737362] env[65522]: DEBUG nova.network.neutron [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Refreshing network info cache for port 3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 695.740604] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "ad074365-1443-481d-990f-479c523219fa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 695.740604] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "ad074365-1443-481d-990f-479c523219fa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 695.740604] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "ad074365-1443-481d-990f-479c523219fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 695.740604] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "ad074365-1443-481d-990f-479c523219fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 695.740871] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "ad074365-1443-481d-990f-479c523219fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 695.743239] env[65522]: INFO nova.compute.manager [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Terminating instance [ 695.745699] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8122a5-ba1d-4e54-b3d6-263ab597dc5c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.754495] env[65522]: DEBUG nova.network.neutron [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Successfully created port: 40d1ea0d-4c4d-4d05-9af8-ba1637b96c97 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 695.827828] env[65522]: DEBUG nova.compute.manager [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 695.836724] env[65522]: DEBUG nova.network.neutron [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Updated VIF entry in instance network info cache for port 97dc7703-dd4a-43dd-8e2d-02f8a31d1a69. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 695.837149] env[65522]: DEBUG nova.network.neutron [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Updating instance_info_cache with network_info: [{"id": "97dc7703-dd4a-43dd-8e2d-02f8a31d1a69", "address": "fa:16:3e:16:68:0f", "network": {"id": "d0ade77e-db62-42ae-9058-c33fcab1d2f9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1207104058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "234a226c969b4965bfa2c411bb4c8d90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97dc7703-dd", "ovs_interfaceid": "97dc7703-dd4a-43dd-8e2d-02f8a31d1a69", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 695.842906] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994fb5b4-dd22-49ef-9534-2aed7ccae305 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.848611] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113697, 'name': CreateVM_Task, 'duration_secs': 0.545985} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.849513] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 695.849991] env[65522]: WARNING neutronclient.v2_0.client [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 695.850468] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.851209] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 695.851209] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 695.853226] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9643cc70-9d4a-4b57-96c5-831bf6796c03 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.858027] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db548ab-d04e-459a-91e1-afd94c76a92f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.864362] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 695.864362] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a89172-b466-4897-2f1c-f3e4a9350676" [ 695.864362] env[65522]: _type = "Task" [ 695.864362] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.881931] env[65522]: DEBUG nova.compute.provider_tree [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.890848] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a89172-b466-4897-2f1c-f3e4a9350676, 'name': SearchDatastore_Task, 'duration_secs': 0.020852} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.893446] env[65522]: DEBUG nova.virt.hardware [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 695.893715] env[65522]: DEBUG nova.virt.hardware [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 695.893823] env[65522]: DEBUG nova.virt.hardware [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 695.894011] env[65522]: DEBUG nova.virt.hardware [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 695.894306] env[65522]: DEBUG nova.virt.hardware [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 695.894306] env[65522]: DEBUG nova.virt.hardware [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 695.894503] env[65522]: DEBUG nova.virt.hardware [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 695.894657] env[65522]: DEBUG nova.virt.hardware [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 695.895170] env[65522]: DEBUG nova.virt.hardware [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 695.895170] env[65522]: DEBUG nova.virt.hardware [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 695.895170] env[65522]: DEBUG nova.virt.hardware [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 695.895527] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 695.896075] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 695.896075] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.896198] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 695.896334] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 695.898046] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea44792-8766-4371-a037-e034e8898843 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.901043] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51e74b7b-be5d-437f-8424-83689b4919fc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.913972] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6a25f5-0f62-41f9-b409-043ffea8af79 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.920110] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 695.920378] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 695.923653] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce429b7a-c870-42e0-8832-a15d6fe720ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.937543] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 695.937543] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528582de-e8f4-8171-9e29-82729d50d003" [ 695.937543] env[65522]: _type = "Task" [ 695.937543] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.946757] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528582de-e8f4-8171-9e29-82729d50d003, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.036278] env[65522]: WARNING openstack [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.036723] env[65522]: WARNING openstack [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.140019] env[65522]: WARNING neutronclient.v2_0.client [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 696.140826] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.141361] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.171222] env[65522]: DEBUG oslo_concurrency.lockutils [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 696.260840] env[65522]: WARNING neutronclient.v2_0.client [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 696.261671] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.262045] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.272924] env[65522]: DEBUG nova.compute.manager [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 696.272924] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 696.273651] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720f9c12-946e-428b-bfab-ddda6928273d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.282505] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 696.282806] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c9e77a2-f692-49fa-be51-3377486c97b4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.291559] env[65522]: DEBUG oslo_vmware.api [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 696.291559] env[65522]: value = "task-5113699" [ 696.291559] env[65522]: _type = "Task" [ 696.291559] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.304282] env[65522]: DEBUG oslo_vmware.api [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113699, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.342307] env[65522]: DEBUG nova.network.neutron [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 696.348985] env[65522]: DEBUG oslo_concurrency.lockutils [req-05a28fa2-64f3-4a93-ae41-a792fc8b0fb8 req-abb38286-30ad-48e5-9ed3-5098158d94e2 service nova] Releasing lock "refresh_cache-0d52b207-ac69-48ce-b3af-7a83a499886e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 696.386430] env[65522]: DEBUG nova.scheduler.client.report [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 696.448885] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528582de-e8f4-8171-9e29-82729d50d003, 'name': SearchDatastore_Task, 'duration_secs': 0.017612} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.450088] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71668541-3c73-4ca1-a38e-4219c76cdb59 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.457984] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 696.457984] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5229e64d-e85a-972e-027f-8dfaee056dff" [ 696.457984] env[65522]: _type = "Task" [ 696.457984] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.467288] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5229e64d-e85a-972e-027f-8dfaee056dff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.802844] env[65522]: DEBUG oslo_vmware.api [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113699, 'name': PowerOffVM_Task, 'duration_secs': 0.344003} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.803149] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 696.803324] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 696.803598] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1445fed1-8197-4a89-bdaf-ab83268ea482 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.841173] env[65522]: WARNING openstack [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.841604] env[65522]: WARNING openstack [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.892366] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.654s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 696.893284] env[65522]: DEBUG nova.compute.manager [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 696.897781] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 696.898186] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 696.902418] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Deleting the datastore file [datastore1] ad074365-1443-481d-990f-479c523219fa {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 696.903655] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.383s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 696.911025] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-21ed86ba-6eb8-4d57-b49a-b5dcd22fb5bd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.918345] env[65522]: DEBUG oslo_vmware.api [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 696.918345] env[65522]: value = "task-5113701" [ 696.918345] env[65522]: _type = "Task" [ 696.918345] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.930460] env[65522]: DEBUG oslo_vmware.api [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113701, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.975878] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5229e64d-e85a-972e-027f-8dfaee056dff, 'name': SearchDatastore_Task, 'duration_secs': 0.017715} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.976242] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 696.976578] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7/f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 696.977563] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-09b2cfef-77c5-4748-bc71-a2d8fa67adea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.987070] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 696.987070] env[65522]: value = "task-5113702" [ 696.987070] env[65522]: _type = "Task" [ 696.987070] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.997186] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113702, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.143196] env[65522]: INFO nova.compute.manager [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Rescuing [ 697.143196] env[65522]: DEBUG oslo_concurrency.lockutils [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "refresh_cache-87a3c63d-794d-44ab-bad6-65c323d72ae7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.406411] env[65522]: DEBUG nova.compute.utils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 697.409546] env[65522]: DEBUG nova.compute.manager [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 697.409546] env[65522]: DEBUG nova.network.neutron [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 697.409943] env[65522]: WARNING neutronclient.v2_0.client [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 697.410116] env[65522]: WARNING neutronclient.v2_0.client [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 697.410712] env[65522]: WARNING openstack [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.411051] env[65522]: WARNING openstack [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.450694] env[65522]: DEBUG oslo_vmware.api [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113701, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.320898} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.451296] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 697.451498] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 697.451670] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 697.451838] env[65522]: INFO nova.compute.manager [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: ad074365-1443-481d-990f-479c523219fa] Took 1.18 seconds to destroy the instance on the hypervisor. [ 697.452102] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 697.452547] env[65522]: DEBUG nova.compute.manager [-] [instance: ad074365-1443-481d-990f-479c523219fa] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 697.452686] env[65522]: DEBUG nova.network.neutron [-] [instance: ad074365-1443-481d-990f-479c523219fa] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 697.454302] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 697.454302] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.454302] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.500525] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113702, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.529370] env[65522]: DEBUG nova.network.neutron [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Successfully updated port: 40d1ea0d-4c4d-4d05-9af8-ba1637b96c97 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 697.661408] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.661478] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.678743] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.679130] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.910454] env[65522]: DEBUG nova.compute.manager [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 697.969042] env[65522]: WARNING neutronclient.v2_0.client [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 697.969165] env[65522]: WARNING openstack [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.969534] env[65522]: WARNING openstack [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.984905] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 57d32fb8-885e-44ba-967b-69f4bd1f5744 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.985633] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance b2d70982-54e7-459c-a0d8-48bf4b6e4345 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.985825] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 1d3a9ada-df46-46af-a7f1-a84c251af4ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.985968] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 6abf58ca-ad6b-4fe8-844a-ca33e62a844d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.986324] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance a09c12be-0aaa-4934-a1b3-5af29feba8a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.988837] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7c6f8218-602d-44f3-8012-de5a96972785 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.988837] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.988837] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 6c8165f5-6769-4954-b9ac-de8551954377 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.988837] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance ad074365-1443-481d-990f-479c523219fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.989322] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance e5d2cc50-4ee6-4e64-9b52-888968a717ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.989322] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance fdf56fa6-ad18-499e-a171-6242abb293bf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.989322] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 0d52b207-ac69-48ce-b3af-7a83a499886e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.989322] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance a8f9f223-023a-48fc-9186-a360a78e27a9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.989768] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 97dc8cac-8241-4912-a3ed-689439ef7ff8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.989768] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance ca964440-5375-4aff-8b45-96fbe829dd16 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.989768] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance cd82f2fd-d657-4ece-8e90-9a47072f5546 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.989768] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 87a3c63d-794d-44ab-bad6-65c323d72ae7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.990162] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 68d8d59b-4fc8-4993-8c9e-31b2329c88dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.990162] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.990162] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance d5e8dd05-dc3c-4831-b4b0-ac100360f3e7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.990162] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance dbf29bd1-780c-4756-a203-4fad05a9d3b9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 697.990381] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance f44db343-7e17-4780-9a67-245992ac1a63 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 698.013300] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113702, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.996497} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.013807] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7/f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 698.014775] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 698.014775] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-298298c7-2d2a-48c1-936a-21da253a248c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.026520] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 698.030451] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 698.030451] env[65522]: value = "task-5113704" [ 698.030451] env[65522]: _type = "Task" [ 698.030451] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.037156] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquiring lock "refresh_cache-dbf29bd1-780c-4756-a203-4fad05a9d3b9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.037156] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquired lock "refresh_cache-dbf29bd1-780c-4756-a203-4fad05a9d3b9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 698.037156] env[65522]: DEBUG nova.network.neutron [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 698.049601] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113704, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.351011] env[65522]: DEBUG nova.policy [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2c898f8d63564df5928a08c323ec1561', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '70696b0fca8842ba8b8eac9cd984504f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 698.402238] env[65522]: WARNING neutronclient.v2_0.client [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 698.404980] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.404980] env[65522]: WARNING openstack [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.430606] env[65522]: WARNING neutronclient.v2_0.client [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 698.431907] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.432515] env[65522]: WARNING openstack [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.447161] env[65522]: DEBUG nova.compute.manager [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 698.449657] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f9dd4bc-fedc-442b-b70f-17a2db4d7466 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.478480] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Acquiring lock "c0c99008-1907-4d3a-b4f6-ee6e916e2c67" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 698.478741] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lock "c0c99008-1907-4d3a-b4f6-ee6e916e2c67" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 698.500281] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 6e10e846-cdb7-458e-b511-18ab1742228e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 698.500578] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 22 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 698.500745] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4800MB phys_disk=100GB used_disk=22GB total_vcpus=48 used_vcpus=22 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '22', 'num_vm_active': '18', 'num_task_None': '17', 'num_os_type_None': '22', 'num_proj_e4fdbd0ed78c41a591ccc0a587313ec8': '2', 'io_workload': '5', 'num_proj_068c2387de8c406194d9b1762c7292a7': '1', 'num_proj_345c026cba1d40c59cab5aa5b477b833': '1', 'num_proj_a47babe780ec4ae7b6c40f4531f09a38': '1', 'num_proj_b1927d35d9154557a33422045152daa6': '1', 'num_proj_fa11b46d9fe144f391233e6eb9c819d7': '1', 'num_proj_7340f053e9784f30be39030d99fe698f': '1', 'num_proj_e2bd38f300424db78568e0b8d9982f27': '1', 'num_task_deleting': '1', 'num_proj_95bca233019f4e2f8bac0dd800392a69': '3', 'num_proj_5f64db4fc2f3412986ef6fd1393e99c8': '1', 'num_proj_234a226c969b4965bfa2c411bb4c8d90': '1', 'num_proj_70696b0fca8842ba8b8eac9cd984504f': '2', 'num_proj_9b8112ba49034e6fabcb3fbbd46edf41': '2', 'num_task_rescuing': '1', 'num_proj_eb9a9e5da1194ab7927f72d21d1484b0': '1', 'num_vm_building': '4', 'num_task_spawning': '3', 'num_proj_6ecccb656b0d4c96b40b200cdcddbad5': '1', 'num_proj_503adb79a1f74f1b98e1f06ec54f1584': '1', 'num_proj_68c8ff7c2b604c939a2124142d1654b8': '1'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 698.552052] env[65522]: WARNING openstack [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.552584] env[65522]: WARNING openstack [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.563036] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113704, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.296621} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.568888] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 698.570332] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aac2c57-17cd-4e58-b38e-c0e6c2063059 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.598970] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Reconfiguring VM instance instance-00000018 to attach disk [datastore1] f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7/f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 698.603118] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9d5d46c-4794-4f2e-a7c0-f805364ff304 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.628403] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 698.628403] env[65522]: value = "task-5113705" [ 698.628403] env[65522]: _type = "Task" [ 698.628403] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.640170] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113705, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.922238] env[65522]: DEBUG nova.compute.manager [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 698.961302] env[65522]: DEBUG nova.virt.hardware [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 698.961736] env[65522]: DEBUG nova.virt.hardware [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 698.961826] env[65522]: DEBUG nova.virt.hardware [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 698.962025] env[65522]: DEBUG nova.virt.hardware [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 698.962191] env[65522]: DEBUG nova.virt.hardware [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 698.962373] env[65522]: DEBUG nova.virt.hardware [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 698.962740] env[65522]: DEBUG nova.virt.hardware [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 698.962807] env[65522]: DEBUG nova.virt.hardware [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 698.963031] env[65522]: DEBUG nova.virt.hardware [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 698.963269] env[65522]: DEBUG nova.virt.hardware [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 698.963489] env[65522]: DEBUG nova.virt.hardware [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 698.964519] env[65522]: INFO nova.compute.manager [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] instance snapshotting [ 698.966717] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a82ecfb3-6aba-4101-ae58-ff00823b9ae9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.971424] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f5bcda-1265-4da6-ad89-e75aa7156cb0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.982208] env[65522]: DEBUG nova.compute.manager [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 699.003551] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3b70a9-09eb-4fb9-8f4c-06f9afcb86a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.011780] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701f694c-1b4c-45be-b7b7-d43c50362f0d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.015621] env[65522]: DEBUG nova.compute.manager [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 699.017412] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc10d86e-2cab-4134-be5a-5deaf6ff18bf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.061165] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3aa154a-3ed5-473e-b210-da0c67c95ec6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.069443] env[65522]: DEBUG nova.network.neutron [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating instance_info_cache with network_info: [{"id": "fb1cd05a-757b-469c-9749-fd1ca4fcb168", "address": "fa:16:3e:b1:9d:ad", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb1cd05a-75", "ovs_interfaceid": "fb1cd05a-757b-469c-9749-fd1ca4fcb168", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 699.074878] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c7db20-abe0-4260-bfa0-8ce0f2afabb4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.110652] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18caf0d7-f3ea-49ff-8625-e1b30f1eaa71 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.119768] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7376a6fc-b2c9-41db-b4cf-83545d9c4ed7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.137409] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 699.150460] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.163391] env[65522]: DEBUG nova.network.neutron [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Successfully created port: 5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 699.248510] env[65522]: DEBUG nova.network.neutron [-] [instance: ad074365-1443-481d-990f-479c523219fa] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 699.312591] env[65522]: DEBUG nova.network.neutron [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 699.432700] env[65522]: DEBUG nova.network.neutron [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Updated VIF entry in instance network info cache for port 60d5d352-4fed-4b0f-9efc-ce13e2066444. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 699.433038] env[65522]: DEBUG nova.network.neutron [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Updating instance_info_cache with network_info: [{"id": "60d5d352-4fed-4b0f-9efc-ce13e2066444", "address": "fa:16:3e:72:1d:e5", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60d5d352-4f", "ovs_interfaceid": "60d5d352-4fed-4b0f-9efc-ce13e2066444", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 699.470853] env[65522]: DEBUG nova.network.neutron [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Updated VIF entry in instance network info cache for port 3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 699.471272] env[65522]: DEBUG nova.network.neutron [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Updating instance_info_cache with network_info: [{"id": "3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6", "address": "fa:16:3e:6f:2f:ba", "network": {"id": "bed09e85-fdc4-4258-9be8-027aa4cf34a6", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-752246656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb9a9e5da1194ab7927f72d21d1484b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de5fcb06-b0d0-467f-86fe-06882165ac31", "external-id": "nsx-vlan-transportzone-595", "segmentation_id": 595, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bbf4b9c-9f", "ovs_interfaceid": "3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 699.532484] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 699.546168] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 699.546871] env[65522]: INFO nova.compute.manager [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] instance snapshotting [ 699.548887] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-325a9f72-f7d7-4de0-9764-1575f8993b5e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.552914] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11f4f71-5f9d-4e8f-b43a-16ed49c28e83 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.558824] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 699.558824] env[65522]: value = "task-5113706" [ 699.558824] env[65522]: _type = "Task" [ 699.558824] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.580986] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Releasing lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.580986] env[65522]: DEBUG nova.compute.manager [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Instance network_info: |[{"id": "fb1cd05a-757b-469c-9749-fd1ca4fcb168", "address": "fa:16:3e:b1:9d:ad", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb1cd05a-75", "ovs_interfaceid": "fb1cd05a-757b-469c-9749-fd1ca4fcb168", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 699.582851] env[65522]: WARNING openstack [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 699.583549] env[65522]: WARNING openstack [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 699.592821] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b1:9d:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fb1cd05a-757b-469c-9749-fd1ca4fcb168', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 699.603318] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Creating folder: Project (503adb79a1f74f1b98e1f06ec54f1584). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 699.604211] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e2e80c9-7c66-4313-9ba1-60b244cf8733 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.612810] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d2bfe098-39ef-4cc8-9568-c05d4bad6f73 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.613953] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113706, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.624104] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Created folder: Project (503adb79a1f74f1b98e1f06ec54f1584) in parent group-v994660. [ 699.624557] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Creating folder: Instances. Parent ref: group-v994730. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 699.625096] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e370d8e7-353b-4751-8d95-2898ef1156e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.636677] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Created folder: Instances in parent group-v994730. [ 699.636897] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 699.637231] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 699.641841] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-22e88c04-3239-4122-8e8b-85bafe82d884 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.668233] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.669922] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 699.669922] env[65522]: value = "task-5113710" [ 699.669922] env[65522]: _type = "Task" [ 699.669922] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.684218] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113710, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.688212] env[65522]: ERROR nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [req-08420c99-cd95-4206-86b9-619b4daa5afd] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9a0d1ac8-ed26-4657-8899-fbceab29db0d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-08420c99-cd95-4206-86b9-619b4daa5afd"}]} [ 699.712784] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Refreshing inventories for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 699.737153] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Updating ProviderTree inventory for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 699.737637] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 699.752237] env[65522]: INFO nova.compute.manager [-] [instance: ad074365-1443-481d-990f-479c523219fa] Took 2.30 seconds to deallocate network for instance. [ 699.760033] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Refreshing aggregate associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, aggregates: None {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 699.792257] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Refreshing trait associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 699.935507] env[65522]: DEBUG oslo_concurrency.lockutils [req-521c9897-e4d8-4046-bd61-38044fe977e7 req-27ad42e9-23a2-46b3-b30b-44660adad0be service nova] Releasing lock "refresh_cache-87a3c63d-794d-44ab-bad6-65c323d72ae7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.936584] env[65522]: DEBUG oslo_concurrency.lockutils [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "refresh_cache-87a3c63d-794d-44ab-bad6-65c323d72ae7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 699.936584] env[65522]: DEBUG nova.network.neutron [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 699.973953] env[65522]: DEBUG oslo_concurrency.lockutils [req-2ae8a8ab-9e0a-4db5-9c40-28e6cbc259d4 req-15251798-86b5-4ec8-95cb-91f1431bf3d4 service nova] Releasing lock "refresh_cache-68d8d59b-4fc8-4993-8c9e-31b2329c88dd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 700.073919] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113706, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.123341] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 700.123784] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1a3586cc-c65d-4894-8c88-aafed64fb0e1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.131632] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 700.131632] env[65522]: value = "task-5113711" [ 700.131632] env[65522]: _type = "Task" [ 700.131632] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.146634] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113711, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.153379] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113705, 'name': ReconfigVM_Task, 'duration_secs': 1.24349} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.153724] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Reconfigured VM instance instance-00000018 to attach disk [datastore1] f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7/f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 700.154476] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-75b757fa-59b7-4550-bb6b-15ee00cc020a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.162793] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 700.162793] env[65522]: value = "task-5113712" [ 700.162793] env[65522]: _type = "Task" [ 700.162793] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.174698] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113712, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.191028] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113710, 'name': CreateVM_Task, 'duration_secs': 0.432208} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.191346] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 700.192453] env[65522]: WARNING neutronclient.v2_0.client [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.192453] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.192453] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 700.192820] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 700.193552] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b51c033f-5545-4950-8f73-97edceef950a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.202852] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 700.202852] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527332d7-c415-6dca-7874-974967d3938b" [ 700.202852] env[65522]: _type = "Task" [ 700.202852] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.219612] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527332d7-c415-6dca-7874-974967d3938b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.266382] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.334785] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8dadcf-816c-4800-869a-4de6613182f1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.346482] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb370d80-fb22-46ae-97cc-da8e540313dc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.382692] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943028a6-c36b-4401-b474-4df6c5b27ab5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.392155] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec5026cc-3d94-4eba-9caf-0243e1ee0e7a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.409255] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 700.438874] env[65522]: WARNING neutronclient.v2_0.client [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.439767] env[65522]: WARNING openstack [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.440259] env[65522]: WARNING openstack [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.574637] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113706, 'name': CreateSnapshot_Task, 'duration_secs': 0.596524} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.574637] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 700.574637] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e11f8d-2e80-4f73-ac9c-e37ddd64aaca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.643971] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113711, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.674461] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113712, 'name': Rename_Task, 'duration_secs': 0.233957} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.675193] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 700.675870] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-34663b0c-b16e-475d-abec-fba06940e20b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.683982] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 700.683982] env[65522]: value = "task-5113713" [ 700.683982] env[65522]: _type = "Task" [ 700.683982] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.693518] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113713, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.714958] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527332d7-c415-6dca-7874-974967d3938b, 'name': SearchDatastore_Task, 'duration_secs': 0.018804} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.715315] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 700.715557] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 700.715800] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.715942] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 700.716222] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 700.716496] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8acadb91-69c0-42cd-b602-99886fbb4f34 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.727544] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 700.727748] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 700.728625] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db8047a3-f567-49ca-b8c2-bc0a281946bf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.735905] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 700.735905] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52399542-72fe-7fa1-ddfb-885bdd913571" [ 700.735905] env[65522]: _type = "Task" [ 700.735905] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.748336] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52399542-72fe-7fa1-ddfb-885bdd913571, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.759025] env[65522]: WARNING neutronclient.v2_0.client [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.759847] env[65522]: WARNING openstack [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.760317] env[65522]: WARNING openstack [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.951731] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Updated inventory for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with generation 35 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 700.951731] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Updating resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d generation from 35 to 36 during operation: update_inventory {{(pid=65522) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 700.951731] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 701.015291] env[65522]: DEBUG nova.compute.manager [req-edfbffc7-c74a-4552-acff-341e2fd2dc65 req-5a844e10-beac-4b25-99d5-323a5bb845a9 service nova] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Received event network-vif-plugged-fb1cd05a-757b-469c-9749-fd1ca4fcb168 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 701.015547] env[65522]: DEBUG oslo_concurrency.lockutils [req-edfbffc7-c74a-4552-acff-341e2fd2dc65 req-5a844e10-beac-4b25-99d5-323a5bb845a9 service nova] Acquiring lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 701.016008] env[65522]: DEBUG oslo_concurrency.lockutils [req-edfbffc7-c74a-4552-acff-341e2fd2dc65 req-5a844e10-beac-4b25-99d5-323a5bb845a9 service nova] Lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 701.016330] env[65522]: DEBUG oslo_concurrency.lockutils [req-edfbffc7-c74a-4552-acff-341e2fd2dc65 req-5a844e10-beac-4b25-99d5-323a5bb845a9 service nova] Lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 701.016520] env[65522]: DEBUG nova.compute.manager [req-edfbffc7-c74a-4552-acff-341e2fd2dc65 req-5a844e10-beac-4b25-99d5-323a5bb845a9 service nova] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] No waiting events found dispatching network-vif-plugged-fb1cd05a-757b-469c-9749-fd1ca4fcb168 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 701.017216] env[65522]: WARNING nova.compute.manager [req-edfbffc7-c74a-4552-acff-341e2fd2dc65 req-5a844e10-beac-4b25-99d5-323a5bb845a9 service nova] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Received unexpected event network-vif-plugged-fb1cd05a-757b-469c-9749-fd1ca4fcb168 for instance with vm_state building and task_state spawning. [ 701.082338] env[65522]: DEBUG nova.network.neutron [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Successfully updated port: 5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 701.096929] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 701.098609] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8c2a5ddb-54e3-4876-b54c-46d703cbb2e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.108501] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 701.108501] env[65522]: value = "task-5113714" [ 701.108501] env[65522]: _type = "Task" [ 701.108501] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.127299] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113714, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.145083] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113711, 'name': CreateSnapshot_Task, 'duration_secs': 0.773299} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.145486] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 701.146408] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be73a04c-1bc0-417b-b4ff-0e8e7059504e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.198350] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113713, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.251428] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52399542-72fe-7fa1-ddfb-885bdd913571, 'name': SearchDatastore_Task, 'duration_secs': 0.016259} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.259380] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8adbcbd3-1f99-4b5d-b6b5-9b71ad74ae9e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.265651] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 701.265651] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526f0ae9-0d79-1077-6a3e-2035b8e3b71c" [ 701.265651] env[65522]: _type = "Task" [ 701.265651] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.276274] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526f0ae9-0d79-1077-6a3e-2035b8e3b71c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.455693] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 701.456017] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.552s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 701.456328] env[65522]: DEBUG oslo_concurrency.lockutils [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.285s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 701.457955] env[65522]: INFO nova.compute.claims [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.506177] env[65522]: DEBUG nova.network.neutron [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Updating instance_info_cache with network_info: [{"id": "40d1ea0d-4c4d-4d05-9af8-ba1637b96c97", "address": "fa:16:3e:fc:a7:6a", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.16", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40d1ea0d-4c", "ovs_interfaceid": "40d1ea0d-4c4d-4d05-9af8-ba1637b96c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 701.555478] env[65522]: WARNING openstack [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 701.556493] env[65522]: WARNING openstack [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.587947] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "refresh_cache-f44db343-7e17-4780-9a67-245992ac1a63" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.587947] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquired lock "refresh_cache-f44db343-7e17-4780-9a67-245992ac1a63" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 701.587947] env[65522]: DEBUG nova.network.neutron [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 701.623562] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113714, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.674052] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 701.676145] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c0885560-9c36-498f-8445-c50d9a08d174 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.681447] env[65522]: DEBUG nova.compute.manager [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Received event network-changed-aca7a84f-d08f-4c79-821e-1f90b00f141b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 701.681650] env[65522]: DEBUG nova.compute.manager [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Refreshing instance network info cache due to event network-changed-aca7a84f-d08f-4c79-821e-1f90b00f141b. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 701.681894] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Acquiring lock "refresh_cache-ad074365-1443-481d-990f-479c523219fa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.682020] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Acquired lock "refresh_cache-ad074365-1443-481d-990f-479c523219fa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 701.683046] env[65522]: DEBUG nova.network.neutron [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Refreshing network info cache for port aca7a84f-d08f-4c79-821e-1f90b00f141b {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 701.686528] env[65522]: WARNING neutronclient.v2_0.client [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 701.687253] env[65522]: WARNING openstack [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 701.687607] env[65522]: WARNING openstack [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.711124] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 701.711124] env[65522]: value = "task-5113715" [ 701.711124] env[65522]: _type = "Task" [ 701.711124] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.715034] env[65522]: DEBUG oslo_vmware.api [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113713, 'name': PowerOnVM_Task, 'duration_secs': 0.539189} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.718532] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 701.718809] env[65522]: INFO nova.compute.manager [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Took 11.81 seconds to spawn the instance on the hypervisor. [ 701.719111] env[65522]: DEBUG nova.compute.manager [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 701.721067] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e520c0fa-18a5-4171-916c-66336eda84cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.731166] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113715, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.782646] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526f0ae9-0d79-1077-6a3e-2035b8e3b71c, 'name': SearchDatastore_Task, 'duration_secs': 0.027545} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.782783] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 701.782971] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] d5e8dd05-dc3c-4831-b4b0-ac100360f3e7/d5e8dd05-dc3c-4831-b4b0-ac100360f3e7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 701.783302] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1922e8e4-dc15-4696-b0cc-0d627093a948 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.791973] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 701.791973] env[65522]: value = "task-5113717" [ 701.791973] env[65522]: _type = "Task" [ 701.791973] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.804602] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113717, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.013979] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Releasing lock "refresh_cache-dbf29bd1-780c-4756-a203-4fad05a9d3b9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 702.014415] env[65522]: DEBUG nova.compute.manager [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Instance network_info: |[{"id": "40d1ea0d-4c4d-4d05-9af8-ba1637b96c97", "address": "fa:16:3e:fc:a7:6a", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.16", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40d1ea0d-4c", "ovs_interfaceid": "40d1ea0d-4c4d-4d05-9af8-ba1637b96c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 702.014885] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:a7:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '40d1ea0d-4c4d-4d05-9af8-ba1637b96c97', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 702.025137] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 702.025323] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 702.025551] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa7f2f5f-65e7-4bdf-95e9-78069392953e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.051471] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 702.051471] env[65522]: value = "task-5113718" [ 702.051471] env[65522]: _type = "Task" [ 702.051471] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.062153] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113718, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.093060] env[65522]: WARNING openstack [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 702.093628] env[65522]: WARNING openstack [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 702.121930] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113714, 'name': CloneVM_Task} progress is 95%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.204930] env[65522]: WARNING neutronclient.v2_0.client [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 702.206453] env[65522]: WARNING openstack [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 702.206453] env[65522]: WARNING openstack [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 702.229546] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113715, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.249078] env[65522]: INFO nova.compute.manager [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Took 30.77 seconds to build instance. [ 702.310708] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113717, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.387221] env[65522]: DEBUG nova.network.neutron [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Updating instance_info_cache with network_info: [{"id": "60d5d352-4fed-4b0f-9efc-ce13e2066444", "address": "fa:16:3e:72:1d:e5", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60d5d352-4f", "ovs_interfaceid": "60d5d352-4fed-4b0f-9efc-ce13e2066444", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 702.428236] env[65522]: DEBUG nova.network.neutron [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 702.448234] env[65522]: DEBUG nova.network.neutron [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 702.568710] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113718, 'name': CreateVM_Task, 'duration_secs': 0.423161} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.573865] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 702.575077] env[65522]: WARNING neutronclient.v2_0.client [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 702.575675] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.575823] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 702.576316] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 702.576557] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c632dc9-47e5-48aa-9d7f-b19768ad559b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.584438] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 702.584438] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526c6af8-21d4-639a-523c-2713df942c27" [ 702.584438] env[65522]: _type = "Task" [ 702.584438] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.601800] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526c6af8-21d4-639a-523c-2713df942c27, 'name': SearchDatastore_Task, 'duration_secs': 0.012347} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.602226] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 702.602477] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 702.602757] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.602944] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 702.603076] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 702.603383] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d767be32-0574-4f61-94b8-87f7edbc53a0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.615217] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 702.615419] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 702.616738] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-516c9c35-2c68-4936-aac8-faa689ca35cb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.623691] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113714, 'name': CloneVM_Task, 'duration_secs': 1.318386} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.624420] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Created linked-clone VM from snapshot [ 702.628040] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef02d42-b52e-4c39-8b40-9bb4b0e2e301 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.633416] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 702.633416] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527ace4f-b35f-d276-be28-fcdc7478b23e" [ 702.633416] env[65522]: _type = "Task" [ 702.633416] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.639973] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Uploading image 1f04bff9-afe0-49b1-9eb7-e369b9348e76 {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 702.653925] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527ace4f-b35f-d276-be28-fcdc7478b23e, 'name': SearchDatastore_Task, 'duration_secs': 0.012659} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.654831] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d52324a8-189b-46f4-a855-92b323aa16f8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.662667] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 702.662667] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5240247b-972b-451a-5be1-fe9380a52c0a" [ 702.662667] env[65522]: _type = "Task" [ 702.662667] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.672020] env[65522]: DEBUG oslo_vmware.rw_handles [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 702.672020] env[65522]: value = "vm-994735" [ 702.672020] env[65522]: _type = "VirtualMachine" [ 702.672020] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 702.672750] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1465a0f2-893b-4ab1-bf14-16a25e6e50f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.679996] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5240247b-972b-451a-5be1-fe9380a52c0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.688587] env[65522]: DEBUG oslo_vmware.rw_handles [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lease: (returnval){ [ 702.688587] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52854220-407d-a9d9-3840-f16bd649f5a5" [ 702.688587] env[65522]: _type = "HttpNfcLease" [ 702.688587] env[65522]: } obtained for exporting VM: (result){ [ 702.688587] env[65522]: value = "vm-994735" [ 702.688587] env[65522]: _type = "VirtualMachine" [ 702.688587] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 702.688587] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the lease: (returnval){ [ 702.688587] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52854220-407d-a9d9-3840-f16bd649f5a5" [ 702.688587] env[65522]: _type = "HttpNfcLease" [ 702.688587] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 702.696211] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 702.696211] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52854220-407d-a9d9-3840-f16bd649f5a5" [ 702.696211] env[65522]: _type = "HttpNfcLease" [ 702.696211] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 702.731084] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113715, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.751491] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1789e2f-0ca4-4d35-bb18-5f78818f5f2f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.284s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 702.804922] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113717, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.806165} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.805381] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] d5e8dd05-dc3c-4831-b4b0-ac100360f3e7/d5e8dd05-dc3c-4831-b4b0-ac100360f3e7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 702.805492] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 702.805765] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c4154a2a-c201-4f8a-ad8d-cc8053413ce1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.814045] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 702.814045] env[65522]: value = "task-5113720" [ 702.814045] env[65522]: _type = "Task" [ 702.814045] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.826877] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113720, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.892221] env[65522]: DEBUG oslo_concurrency.lockutils [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "refresh_cache-87a3c63d-794d-44ab-bad6-65c323d72ae7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 702.900946] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad39c458-9c60-4989-82bf-ec66a5f9522c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.909697] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8d7ed5-2974-49af-9538-d99a1157ae83 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.960299] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a416b07a-6890-47ed-bffc-063d42cb9db1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.970063] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77607e65-42ec-41ab-86c9-444ba84ad7ad {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.976887] env[65522]: WARNING openstack [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 702.977360] env[65522]: WARNING openstack [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 702.997587] env[65522]: DEBUG nova.compute.provider_tree [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 703.180575] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5240247b-972b-451a-5be1-fe9380a52c0a, 'name': SearchDatastore_Task, 'duration_secs': 0.012356} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.180799] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 703.181073] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] dbf29bd1-780c-4756-a203-4fad05a9d3b9/dbf29bd1-780c-4756-a203-4fad05a9d3b9.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 703.181513] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-33a294ea-2d55-44e5-8ba9-b4b63370b871 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.189668] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 703.189668] env[65522]: value = "task-5113721" [ 703.189668] env[65522]: _type = "Task" [ 703.189668] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.201907] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 703.201907] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52854220-407d-a9d9-3840-f16bd649f5a5" [ 703.201907] env[65522]: _type = "HttpNfcLease" [ 703.201907] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 703.206393] env[65522]: DEBUG oslo_vmware.rw_handles [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 703.206393] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52854220-407d-a9d9-3840-f16bd649f5a5" [ 703.206393] env[65522]: _type = "HttpNfcLease" [ 703.206393] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 703.206393] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113721, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.207255] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c4de9f1-ef55-4e91-be73-80db75d2b4a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.213438] env[65522]: DEBUG nova.network.neutron [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 703.224656] env[65522]: DEBUG oslo_vmware.rw_handles [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ebbd16-f407-eec0-0536-1f815ffa9ca5/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 703.224656] env[65522]: DEBUG oslo_vmware.rw_handles [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ebbd16-f407-eec0-0536-1f815ffa9ca5/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 703.297227] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113715, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.326484] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113720, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079333} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.326484] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 703.326484] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-790fe834-1c9e-493e-b5ed-eb655b89b680 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.353185] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Reconfiguring VM instance instance-00000019 to attach disk [datastore1] d5e8dd05-dc3c-4831-b4b0-ac100360f3e7/d5e8dd05-dc3c-4831-b4b0-ac100360f3e7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 703.353493] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf7150ff-94ed-40dd-a614-0901c960bb64 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.377173] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 703.377173] env[65522]: value = "task-5113722" [ 703.377173] env[65522]: _type = "Task" [ 703.377173] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.388506] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113722, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.398865] env[65522]: WARNING neutronclient.v2_0.client [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.399620] env[65522]: WARNING openstack [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.399950] env[65522]: WARNING openstack [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.423047] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a9602d02-c442-4282-8b75-90a95ecdb8cd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.530574] env[65522]: DEBUG nova.network.neutron [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Updating instance_info_cache with network_info: [{"id": "5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e", "address": "fa:16:3e:97:28:bd", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.112", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e8b1f95-61", "ovs_interfaceid": "5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 703.551981] env[65522]: DEBUG nova.scheduler.client.report [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Updated inventory for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with generation 36 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 703.553574] env[65522]: DEBUG nova.compute.provider_tree [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Updating resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d generation from 36 to 37 during operation: update_inventory {{(pid=65522) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 703.553574] env[65522]: DEBUG nova.compute.provider_tree [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 703.705341] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113721, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.718174] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Releasing lock "refresh_cache-ad074365-1443-481d-990f-479c523219fa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 703.718570] env[65522]: DEBUG nova.compute.manager [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Received event network-vif-plugged-189bde05-0b1d-4c6a-be88-2453345dd731 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 703.720200] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Acquiring lock "f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 703.721021] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Lock "f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 703.721196] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Lock "f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 703.721463] env[65522]: DEBUG nova.compute.manager [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] No waiting events found dispatching network-vif-plugged-189bde05-0b1d-4c6a-be88-2453345dd731 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 703.721809] env[65522]: WARNING nova.compute.manager [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Received unexpected event network-vif-plugged-189bde05-0b1d-4c6a-be88-2453345dd731 for instance with vm_state building and task_state spawning. [ 703.722056] env[65522]: DEBUG nova.compute.manager [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Received event network-changed-189bde05-0b1d-4c6a-be88-2453345dd731 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 703.722284] env[65522]: DEBUG nova.compute.manager [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Refreshing instance network info cache due to event network-changed-189bde05-0b1d-4c6a-be88-2453345dd731. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 703.722584] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Acquiring lock "refresh_cache-f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.722808] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Acquired lock "refresh_cache-f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 703.723065] env[65522]: DEBUG nova.network.neutron [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Refreshing network info cache for port 189bde05-0b1d-4c6a-be88-2453345dd731 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 703.743323] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113715, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.897086] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113722, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.033803] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Releasing lock "refresh_cache-f44db343-7e17-4780-9a67-245992ac1a63" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.034336] env[65522]: DEBUG nova.compute.manager [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Instance network_info: |[{"id": "5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e", "address": "fa:16:3e:97:28:bd", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.112", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e8b1f95-61", "ovs_interfaceid": "5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 704.036992] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:97:28:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 704.046740] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 704.046740] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 704.047029] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a672f8b8-4cf6-4d22-a79a-55c089f944bc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.065612] env[65522]: DEBUG oslo_concurrency.lockutils [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 704.066372] env[65522]: DEBUG nova.compute.manager [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 704.069360] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.537s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 704.074280] env[65522]: INFO nova.compute.claims [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 704.089771] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 704.089771] env[65522]: value = "task-5113723" [ 704.089771] env[65522]: _type = "Task" [ 704.089771] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.105583] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113723, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.210719] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113721, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.235508] env[65522]: WARNING neutronclient.v2_0.client [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 704.236262] env[65522]: WARNING openstack [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 704.236742] env[65522]: WARNING openstack [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 704.253317] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113715, 'name': CloneVM_Task} progress is 95%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.391825] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113722, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.484166] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 704.484478] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1030794b-61c6-46d5-b9f3-e65103789588 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.493460] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 704.493460] env[65522]: value = "task-5113724" [ 704.493460] env[65522]: _type = "Task" [ 704.493460] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.506191] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113724, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.580508] env[65522]: DEBUG nova.compute.utils [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 704.585242] env[65522]: DEBUG nova.compute.manager [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Not allocating networking since 'none' was specified. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 704.603763] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113723, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.709237] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113721, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.092017} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.709797] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] dbf29bd1-780c-4756-a203-4fad05a9d3b9/dbf29bd1-780c-4756-a203-4fad05a9d3b9.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 704.711146] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 704.711653] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e23d09a3-4943-4947-ae14-7c791e00fa2e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.726535] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 704.726535] env[65522]: value = "task-5113725" [ 704.726535] env[65522]: _type = "Task" [ 704.726535] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.745186] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113725, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.750171] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113715, 'name': CloneVM_Task, 'duration_secs': 2.724171} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.750273] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Created linked-clone VM from snapshot [ 704.751105] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd5aa3a-3240-4137-9707-06aa10ef32a0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.769044] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Uploading image 4cd7fa95-e6d9-457c-8cf1-e3881f933468 {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 704.800861] env[65522]: DEBUG oslo_vmware.rw_handles [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 704.800861] env[65522]: value = "vm-994736" [ 704.800861] env[65522]: _type = "VirtualMachine" [ 704.800861] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 704.802452] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d933e41b-be8c-4e5b-8525-5ceb57a36308 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.811167] env[65522]: DEBUG oslo_vmware.rw_handles [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lease: (returnval){ [ 704.811167] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5209e473-5daf-4a57-66e1-16e5ef10985d" [ 704.811167] env[65522]: _type = "HttpNfcLease" [ 704.811167] env[65522]: } obtained for exporting VM: (result){ [ 704.811167] env[65522]: value = "vm-994736" [ 704.811167] env[65522]: _type = "VirtualMachine" [ 704.811167] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 704.811566] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the lease: (returnval){ [ 704.811566] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5209e473-5daf-4a57-66e1-16e5ef10985d" [ 704.811566] env[65522]: _type = "HttpNfcLease" [ 704.811566] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 704.820697] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 704.820697] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5209e473-5daf-4a57-66e1-16e5ef10985d" [ 704.820697] env[65522]: _type = "HttpNfcLease" [ 704.820697] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 704.893439] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113722, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.005554] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113724, 'name': PowerOffVM_Task, 'duration_secs': 0.274445} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.005974] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 705.007034] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced28567-c4ea-4cb6-b046-430976feb236 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.039642] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0abe642-4ef1-4959-a8c3-217718cd5992 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.062211] env[65522]: WARNING openstack [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 705.062648] env[65522]: WARNING openstack [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 705.087120] env[65522]: DEBUG nova.compute.manager [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 705.095026] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 705.099057] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-602643e6-5c45-4266-9a4a-5526d9d32341 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.113123] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113723, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.114546] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 705.114546] env[65522]: value = "task-5113727" [ 705.114546] env[65522]: _type = "Task" [ 705.114546] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.125318] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 705.125716] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 705.126029] env[65522]: DEBUG oslo_concurrency.lockutils [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.126307] env[65522]: DEBUG oslo_concurrency.lockutils [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 705.126546] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 705.130427] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fecfca0d-280a-4d97-8a29-97ddfb53dc06 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.156024] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 705.156024] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 705.156024] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70250553-0f36-4946-a52d-f23c54ed70d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.167348] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 705.167348] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52689733-adfe-d0e0-0eb8-6682a87b0b14" [ 705.167348] env[65522]: _type = "Task" [ 705.167348] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.181276] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52689733-adfe-d0e0-0eb8-6682a87b0b14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.238385] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113725, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.163043} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.238719] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 705.240867] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab141714-918b-4f09-9960-2858f364e92d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.273721] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Reconfiguring VM instance instance-0000001a to attach disk [datastore1] dbf29bd1-780c-4756-a203-4fad05a9d3b9/dbf29bd1-780c-4756-a203-4fad05a9d3b9.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 705.277461] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38ff4f86-8f7c-417d-9547-4c5fa1d15221 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.294890] env[65522]: WARNING neutronclient.v2_0.client [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 705.295721] env[65522]: WARNING openstack [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 705.296251] env[65522]: WARNING openstack [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 705.312657] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 705.312657] env[65522]: value = "task-5113728" [ 705.312657] env[65522]: _type = "Task" [ 705.312657] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.329970] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 705.329970] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5209e473-5daf-4a57-66e1-16e5ef10985d" [ 705.329970] env[65522]: _type = "HttpNfcLease" [ 705.329970] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 705.330237] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113728, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.334836] env[65522]: DEBUG oslo_vmware.rw_handles [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 705.334836] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5209e473-5daf-4a57-66e1-16e5ef10985d" [ 705.334836] env[65522]: _type = "HttpNfcLease" [ 705.334836] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 705.338459] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69cf4ee-e495-4ccb-a8a7-89836de963c2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.352374] env[65522]: DEBUG oslo_vmware.rw_handles [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52037107-5b13-9e63-8be1-bcc379b5dba5/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 705.352885] env[65522]: DEBUG oslo_vmware.rw_handles [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52037107-5b13-9e63-8be1-bcc379b5dba5/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 705.438653] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113722, 'name': ReconfigVM_Task, 'duration_secs': 1.55356} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.438653] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Reconfigured VM instance instance-00000019 to attach disk [datastore1] d5e8dd05-dc3c-4831-b4b0-ac100360f3e7/d5e8dd05-dc3c-4831-b4b0-ac100360f3e7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 705.439284] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ec6eaf66-7f42-42c6-b3b2-a59c1a93923b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.449237] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 705.449237] env[65522]: value = "task-5113729" [ 705.449237] env[65522]: _type = "Task" [ 705.449237] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.465915] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113729, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.478271] env[65522]: DEBUG nova.network.neutron [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Updated VIF entry in instance network info cache for port 189bde05-0b1d-4c6a-be88-2453345dd731. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 705.478696] env[65522]: DEBUG nova.network.neutron [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Updating instance_info_cache with network_info: [{"id": "189bde05-0b1d-4c6a-be88-2453345dd731", "address": "fa:16:3e:76:53:d6", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap189bde05-0b", "ovs_interfaceid": "189bde05-0b1d-4c6a-be88-2453345dd731", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 705.515759] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0a359bfd-f4a3-4598-842c-6afc0dcfe902 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.626154] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113723, 'name': CreateVM_Task, 'duration_secs': 1.149267} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.626154] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 705.626848] env[65522]: WARNING neutronclient.v2_0.client [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 705.627418] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.627564] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 705.629735] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 705.629735] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f32ca0f3-60ef-4c3c-8ebd-658c5e1f61d1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.640222] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 705.640222] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528fbeb6-90fd-404c-02fd-155565126b3d" [ 705.640222] env[65522]: _type = "Task" [ 705.640222] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.660112] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528fbeb6-90fd-404c-02fd-155565126b3d, 'name': SearchDatastore_Task, 'duration_secs': 0.01823} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.663532] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 705.663532] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 705.663862] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.663900] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 705.666016] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 705.666016] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eccb36c5-8c5e-4abf-8923-36bfe40f6af6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.678554] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 705.678774] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 705.680486] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f84b58c8-d14f-44a0-b26e-f890a2bc825b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.686875] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52689733-adfe-d0e0-0eb8-6682a87b0b14, 'name': SearchDatastore_Task, 'duration_secs': 0.021454} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.690902] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b6b042b-a360-460d-8265-5178578e52d5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.696925] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 705.696925] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5271511f-56f9-8e2d-288f-35d06820e731" [ 705.696925] env[65522]: _type = "Task" [ 705.696925] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.706421] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 705.706421] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f1cb1d-688d-8a17-f717-7ebd8405c139" [ 705.706421] env[65522]: _type = "Task" [ 705.706421] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.707770] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5271511f-56f9-8e2d-288f-35d06820e731, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.723679] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f1cb1d-688d-8a17-f717-7ebd8405c139, 'name': SearchDatastore_Task, 'duration_secs': 0.020213} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.724593] env[65522]: DEBUG oslo_concurrency.lockutils [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 705.724970] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 87a3c63d-794d-44ab-bad6-65c323d72ae7/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk. {{(pid=65522) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 705.725303] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cc89679e-75ff-4c58-a577-719de593c875 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.737808] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 705.737808] env[65522]: value = "task-5113730" [ 705.737808] env[65522]: _type = "Task" [ 705.737808] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.752796] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113730, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.802257] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe8ba16-afdb-4e72-997b-92d454b104c2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.813056] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba0cc13-1777-4c4a-b667-076d4885928b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.857232] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c0a925-5ed8-4e40-9d07-640e8d86c71d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.865774] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113728, 'name': ReconfigVM_Task, 'duration_secs': 0.390881} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.866849] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Reconfigured VM instance instance-0000001a to attach disk [datastore1] dbf29bd1-780c-4756-a203-4fad05a9d3b9/dbf29bd1-780c-4756-a203-4fad05a9d3b9.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 705.867770] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a6ffdce-4702-4167-9b9c-0f8c6520cbfa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.880122] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71732165-b06a-4154-b52e-ad6e2ae96363 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.885565] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 705.885565] env[65522]: value = "task-5113731" [ 705.885565] env[65522]: _type = "Task" [ 705.885565] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.901243] env[65522]: DEBUG nova.compute.provider_tree [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.905921] env[65522]: DEBUG nova.compute.manager [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Received event network-changed-fb1cd05a-757b-469c-9749-fd1ca4fcb168 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 705.906173] env[65522]: DEBUG nova.compute.manager [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Refreshing instance network info cache due to event network-changed-fb1cd05a-757b-469c-9749-fd1ca4fcb168. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 705.906767] env[65522]: DEBUG oslo_concurrency.lockutils [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] Acquiring lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.906994] env[65522]: DEBUG oslo_concurrency.lockutils [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] Acquired lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 705.907276] env[65522]: DEBUG nova.network.neutron [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Refreshing network info cache for port fb1cd05a-757b-469c-9749-fd1ca4fcb168 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 705.914337] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113731, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.960711] env[65522]: DEBUG nova.compute.manager [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Received event network-vif-plugged-40d1ea0d-4c4d-4d05-9af8-ba1637b96c97 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 705.960976] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Acquiring lock "dbf29bd1-780c-4756-a203-4fad05a9d3b9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.961297] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Lock "dbf29bd1-780c-4756-a203-4fad05a9d3b9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 705.962643] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Lock "dbf29bd1-780c-4756-a203-4fad05a9d3b9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 705.962793] env[65522]: DEBUG nova.compute.manager [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] No waiting events found dispatching network-vif-plugged-40d1ea0d-4c4d-4d05-9af8-ba1637b96c97 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 705.962971] env[65522]: WARNING nova.compute.manager [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Received unexpected event network-vif-plugged-40d1ea0d-4c4d-4d05-9af8-ba1637b96c97 for instance with vm_state building and task_state spawning. [ 705.963162] env[65522]: DEBUG nova.compute.manager [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Received event network-changed-40d1ea0d-4c4d-4d05-9af8-ba1637b96c97 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 705.963296] env[65522]: DEBUG nova.compute.manager [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Refreshing instance network info cache due to event network-changed-40d1ea0d-4c4d-4d05-9af8-ba1637b96c97. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 705.963487] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Acquiring lock "refresh_cache-dbf29bd1-780c-4756-a203-4fad05a9d3b9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.963670] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Acquired lock "refresh_cache-dbf29bd1-780c-4756-a203-4fad05a9d3b9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 705.963766] env[65522]: DEBUG nova.network.neutron [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Refreshing network info cache for port 40d1ea0d-4c4d-4d05-9af8-ba1637b96c97 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 705.971980] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113729, 'name': Rename_Task, 'duration_secs': 0.189505} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.972599] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 705.972814] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4bb649c9-d665-4953-a084-44a5d2009563 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.981606] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3694e85-7aee-496c-b24e-f40be673b2d7 req-e0c0355f-812a-4e6a-ad10-9ecaddf22f9d service nova] Releasing lock "refresh_cache-f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 705.985207] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 705.985207] env[65522]: value = "task-5113732" [ 705.985207] env[65522]: _type = "Task" [ 705.985207] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.999042] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113732, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.116707] env[65522]: DEBUG nova.compute.manager [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 706.211334] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5271511f-56f9-8e2d-288f-35d06820e731, 'name': SearchDatastore_Task, 'duration_secs': 0.015387} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.213172] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1da21401-1f58-4b15-b658-f01ccada0ca9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.223543] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 706.223543] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520f0c88-13cf-fd5d-7ef2-346bbceec7f5" [ 706.223543] env[65522]: _type = "Task" [ 706.223543] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.241794] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "75c8848b-aea9-43f1-8697-9224050d1fef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 706.242129] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "75c8848b-aea9-43f1-8697-9224050d1fef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.244403] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520f0c88-13cf-fd5d-7ef2-346bbceec7f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.257639] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113730, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.406905] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113731, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.410770] env[65522]: DEBUG nova.scheduler.client.report [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 706.416428] env[65522]: WARNING neutronclient.v2_0.client [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 706.419813] env[65522]: WARNING openstack [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 706.419813] env[65522]: WARNING openstack [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 706.471037] env[65522]: WARNING neutronclient.v2_0.client [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 706.471037] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 706.471037] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 706.503945] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113732, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.739285] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520f0c88-13cf-fd5d-7ef2-346bbceec7f5, 'name': SearchDatastore_Task, 'duration_secs': 0.032753} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.739581] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 706.739984] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] f44db343-7e17-4780-9a67-245992ac1a63/f44db343-7e17-4780-9a67-245992ac1a63.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 706.740137] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aaf46285-4541-4f7c-b701-c0e7409d8607 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.753783] env[65522]: DEBUG nova.compute.manager [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 706.756537] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 706.756537] env[65522]: value = "task-5113733" [ 706.756537] env[65522]: _type = "Task" [ 706.756537] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.766402] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113730, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.845835} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.767455] env[65522]: INFO nova.virt.vmwareapi.ds_util [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 87a3c63d-794d-44ab-bad6-65c323d72ae7/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk. [ 706.768739] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb1f73f6-b487-4179-baed-0f14771352a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.777587] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113733, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.810044] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Reconfiguring VM instance instance-00000016 to attach disk [datastore2] 87a3c63d-794d-44ab-bad6-65c323d72ae7/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 706.810044] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-572e177c-32a4-4545-897c-50b5b7c02f78 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.835929] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 706.835929] env[65522]: value = "task-5113734" [ 706.835929] env[65522]: _type = "Task" [ 706.835929] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.846940] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113734, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.901017] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113731, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.935603] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.866s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.936310] env[65522]: DEBUG nova.compute.manager [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 706.939932] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.673s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.941100] env[65522]: DEBUG nova.objects.instance [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lazy-loading 'resources' on Instance uuid ad074365-1443-481d-990f-479c523219fa {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 707.007100] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113732, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.107315] env[65522]: WARNING openstack [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.109022] env[65522]: WARNING openstack [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.126965] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.127476] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.193295] env[65522]: WARNING neutronclient.v2_0.client [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 707.194984] env[65522]: WARNING openstack [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.195661] env[65522]: WARNING openstack [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.217041] env[65522]: WARNING neutronclient.v2_0.client [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 707.217781] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.218440] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.286582] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113733, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.293416] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 707.356855] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113734, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.369739] env[65522]: DEBUG nova.network.neutron [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updated VIF entry in instance network info cache for port fb1cd05a-757b-469c-9749-fd1ca4fcb168. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 707.369739] env[65522]: DEBUG nova.network.neutron [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating instance_info_cache with network_info: [{"id": "fb1cd05a-757b-469c-9749-fd1ca4fcb168", "address": "fa:16:3e:b1:9d:ad", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb1cd05a-75", "ovs_interfaceid": "fb1cd05a-757b-469c-9749-fd1ca4fcb168", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 707.402529] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113731, 'name': Rename_Task, 'duration_secs': 1.14619} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.402895] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 707.403368] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-444c21f4-bbba-4100-82c6-eb69b29d6409 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.414952] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 707.414952] env[65522]: value = "task-5113735" [ 707.414952] env[65522]: _type = "Task" [ 707.414952] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.436303] env[65522]: DEBUG nova.network.neutron [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Updated VIF entry in instance network info cache for port 40d1ea0d-4c4d-4d05-9af8-ba1637b96c97. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 707.436793] env[65522]: DEBUG nova.network.neutron [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Updating instance_info_cache with network_info: [{"id": "40d1ea0d-4c4d-4d05-9af8-ba1637b96c97", "address": "fa:16:3e:fc:a7:6a", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.16", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40d1ea0d-4c", "ovs_interfaceid": "40d1ea0d-4c4d-4d05-9af8-ba1637b96c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 707.441988] env[65522]: DEBUG nova.compute.utils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 707.443850] env[65522]: DEBUG nova.compute.manager [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 707.444141] env[65522]: DEBUG nova.network.neutron [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 707.444653] env[65522]: WARNING neutronclient.v2_0.client [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 707.444941] env[65522]: WARNING neutronclient.v2_0.client [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 707.445627] env[65522]: WARNING openstack [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.446057] env[65522]: WARNING openstack [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.508546] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113732, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.552970] env[65522]: DEBUG nova.policy [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69075f9a26694ad584114b219ee05e7b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7f98c0ae38740df877c198e2cb611a0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 707.777597] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113733, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.943664} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.783034] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] f44db343-7e17-4780-9a67-245992ac1a63/f44db343-7e17-4780-9a67-245992ac1a63.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 707.783358] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 707.785143] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8d7510b-deaf-4f79-aeb4-e084b32167cb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.793522] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 707.793522] env[65522]: value = "task-5113736" [ 707.793522] env[65522]: _type = "Task" [ 707.793522] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.810833] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113736, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.863061] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113734, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.877181] env[65522]: DEBUG oslo_concurrency.lockutils [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] Releasing lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 707.878087] env[65522]: DEBUG nova.compute.manager [req-6d891466-f86b-42ce-95c2-2cd364825efe req-fac82b56-3f36-45ff-af35-a7eca9c09944 service nova] [instance: ad074365-1443-481d-990f-479c523219fa] Received event network-vif-deleted-aca7a84f-d08f-4c79-821e-1f90b00f141b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 707.931277] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113735, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.942714] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Releasing lock "refresh_cache-dbf29bd1-780c-4756-a203-4fad05a9d3b9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 707.943189] env[65522]: DEBUG nova.compute.manager [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Received event network-vif-plugged-5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 707.943484] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Acquiring lock "f44db343-7e17-4780-9a67-245992ac1a63-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 707.943817] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Lock "f44db343-7e17-4780-9a67-245992ac1a63-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 707.944054] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Lock "f44db343-7e17-4780-9a67-245992ac1a63-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 707.944576] env[65522]: DEBUG nova.compute.manager [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: f44db343-7e17-4780-9a67-245992ac1a63] No waiting events found dispatching network-vif-plugged-5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 707.945130] env[65522]: WARNING nova.compute.manager [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Received unexpected event network-vif-plugged-5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e for instance with vm_state building and task_state spawning. [ 707.945506] env[65522]: DEBUG nova.compute.manager [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Received event network-changed-5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 707.945808] env[65522]: DEBUG nova.compute.manager [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Refreshing instance network info cache due to event network-changed-5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 707.946415] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Acquiring lock "refresh_cache-f44db343-7e17-4780-9a67-245992ac1a63" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.946606] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Acquired lock "refresh_cache-f44db343-7e17-4780-9a67-245992ac1a63" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 707.946826] env[65522]: DEBUG nova.network.neutron [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Refreshing network info cache for port 5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 707.959876] env[65522]: DEBUG nova.compute.manager [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 708.004751] env[65522]: DEBUG oslo_vmware.api [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113732, 'name': PowerOnVM_Task, 'duration_secs': 1.630112} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.009074] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 708.009663] env[65522]: INFO nova.compute.manager [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Took 16.02 seconds to spawn the instance on the hypervisor. [ 708.009663] env[65522]: DEBUG nova.compute.manager [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 708.011087] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a661f527-185d-415d-bcb4-3c69ec6371e9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.028179] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed21c72-5e70-4720-b2b8-ecb898897575 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.036373] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7eff87a-1084-494c-89a2-69b4facb5fe8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.041659] env[65522]: DEBUG nova.network.neutron [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Successfully created port: bb734aa2-6103-4321-b1bd-f9f5a80296ed {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 708.079920] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-527e2922-b192-4d73-9726-23e4b525d5a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.090904] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5833949d-937a-48ba-8465-d6aa612c6753 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.108116] env[65522]: DEBUG nova.compute.provider_tree [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.310138] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113736, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083743} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.310440] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 708.311423] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34860cef-cba6-438a-a107-eaf0a4a29ed1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.337611] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Reconfiguring VM instance instance-0000001b to attach disk [datastore1] f44db343-7e17-4780-9a67-245992ac1a63/f44db343-7e17-4780-9a67-245992ac1a63.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 708.337961] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94ca966e-02a3-4f97-9824-0ec5ce87af3d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.366484] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113734, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.368053] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 708.368053] env[65522]: value = "task-5113737" [ 708.368053] env[65522]: _type = "Task" [ 708.368053] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.385840] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113737, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.435345] env[65522]: DEBUG oslo_vmware.api [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113735, 'name': PowerOnVM_Task, 'duration_secs': 0.805258} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.435345] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 708.435345] env[65522]: INFO nova.compute.manager [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Took 12.61 seconds to spawn the instance on the hypervisor. [ 708.435345] env[65522]: DEBUG nova.compute.manager [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 708.436566] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36aa4f9e-a3f9-4371-b97a-91f0256c8eb6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.457216] env[65522]: WARNING neutronclient.v2_0.client [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 708.457216] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 708.457216] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 708.537927] env[65522]: INFO nova.compute.manager [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Took 35.61 seconds to build instance. [ 708.612699] env[65522]: DEBUG nova.scheduler.client.report [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 708.709483] env[65522]: DEBUG oslo_concurrency.lockutils [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Acquiring lock "42a08cfb-d865-4967-a086-370a2ca98b7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 708.710117] env[65522]: DEBUG oslo_concurrency.lockutils [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Lock "42a08cfb-d865-4967-a086-370a2ca98b7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 708.867884] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113734, 'name': ReconfigVM_Task, 'duration_secs': 1.693121} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.868910] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Reconfigured VM instance instance-00000016 to attach disk [datastore2] 87a3c63d-794d-44ab-bad6-65c323d72ae7/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 708.871070] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b061d0-a94e-4819-a24f-e278582ce31c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.926460] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113737, 'name': ReconfigVM_Task, 'duration_secs': 0.519691} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.926460] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dca232c0-2db2-4cf5-95e2-dcbd55cb9808 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.938221] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Reconfigured VM instance instance-0000001b to attach disk [datastore1] f44db343-7e17-4780-9a67-245992ac1a63/f44db343-7e17-4780-9a67-245992ac1a63.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 708.938944] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2215c9eb-0f19-43a7-bc53-bea99d1dc3aa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.948338] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 708.948338] env[65522]: value = "task-5113739" [ 708.948338] env[65522]: _type = "Task" [ 708.948338] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.949959] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 708.949959] env[65522]: value = "task-5113738" [ 708.949959] env[65522]: _type = "Task" [ 708.949959] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.963854] env[65522]: INFO nova.compute.manager [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Took 31.96 seconds to build instance. [ 708.972901] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113739, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.973883] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.979524] env[65522]: DEBUG nova.compute.manager [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 709.040617] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f16602b-f324-4437-9a33-ac179460ba2d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.158s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.118714] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.179s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.121477] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.828s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.127034] env[65522]: INFO nova.compute.claims [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 709.155131] env[65522]: INFO nova.scheduler.client.report [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Deleted allocations for instance ad074365-1443-481d-990f-479c523219fa [ 709.212929] env[65522]: DEBUG nova.compute.manager [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 709.471021] env[65522]: DEBUG oslo_concurrency.lockutils [None req-41a6e3d9-276f-48ed-9ddc-eddf30498c9a tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lock "dbf29bd1-780c-4756-a203-4fad05a9d3b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.489s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.471453] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113738, 'name': ReconfigVM_Task, 'duration_secs': 0.30271} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.471785] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113739, 'name': Rename_Task, 'duration_secs': 0.253714} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.472096] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 709.472407] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 709.472702] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-83d6271c-ecd0-40d1-b6b8-f28c836df68d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.474596] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46c97a70-a632-463b-87b1-f8b168cdb74b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.483903] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 709.483903] env[65522]: value = "task-5113740" [ 709.483903] env[65522]: _type = "Task" [ 709.483903] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.489343] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 709.489343] env[65522]: value = "task-5113741" [ 709.489343] env[65522]: _type = "Task" [ 709.489343] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.497038] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113740, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.506883] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113741, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.667771] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0666cff0-5ad1-474f-9a7e-be52c1551093 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "ad074365-1443-481d-990f-479c523219fa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.927s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.750497] env[65522]: DEBUG oslo_concurrency.lockutils [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 709.917522] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 709.918064] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 709.943071] env[65522]: DEBUG nova.network.neutron [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Successfully updated port: bb734aa2-6103-4321-b1bd-f9f5a80296ed {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 710.019493] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113741, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.019981] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113740, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.055503] env[65522]: WARNING neutronclient.v2_0.client [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 710.056317] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 710.056710] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 710.173499] env[65522]: DEBUG nova.network.neutron [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Updated VIF entry in instance network info cache for port 5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 710.177028] env[65522]: DEBUG nova.network.neutron [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Updating instance_info_cache with network_info: [{"id": "5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e", "address": "fa:16:3e:97:28:bd", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.112", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e8b1f95-61", "ovs_interfaceid": "5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 710.446940] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Acquiring lock "refresh_cache-c0c99008-1907-4d3a-b4f6-ee6e916e2c67" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.446940] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Acquired lock "refresh_cache-c0c99008-1907-4d3a-b4f6-ee6e916e2c67" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 710.448290] env[65522]: DEBUG nova.network.neutron [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 710.516296] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113741, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.516723] env[65522]: DEBUG oslo_vmware.api [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113740, 'name': PowerOnVM_Task, 'duration_secs': 0.928498} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.517036] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 710.517351] env[65522]: INFO nova.compute.manager [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Took 11.59 seconds to spawn the instance on the hypervisor. [ 710.517569] env[65522]: DEBUG nova.compute.manager [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 710.518617] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffcbc05e-5a3a-4906-9125-702112c356ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.653701] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c29b0f7-35b2-4089-a5b6-aec227cdb89c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.665623] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63b25cc-c6e0-4071-8754-39b0715e33e3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.708602] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Releasing lock "refresh_cache-f44db343-7e17-4780-9a67-245992ac1a63" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 710.709975] env[65522]: DEBUG nova.compute.manager [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Received event network-changed-7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 710.709975] env[65522]: DEBUG nova.compute.manager [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Refreshing instance network info cache due to event network-changed-7d49f270-6cbd-4767-91fc-f899c46564c9. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 710.709975] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Acquiring lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.709975] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Acquired lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 710.709975] env[65522]: DEBUG nova.network.neutron [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Refreshing network info cache for port 7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 710.712977] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfbe8259-a145-424a-83ee-3459e2470ee0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.725788] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc721c2-4328-4cc5-96d5-3f21bca2c516 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.748485] env[65522]: DEBUG nova.compute.provider_tree [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.951762] env[65522]: WARNING openstack [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 710.955029] env[65522]: WARNING openstack [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 711.010537] env[65522]: DEBUG oslo_vmware.api [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113741, 'name': PowerOnVM_Task, 'duration_secs': 1.257001} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.011910] env[65522]: DEBUG nova.network.neutron [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 711.014198] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 711.021042] env[65522]: DEBUG nova.compute.manager [None req-44e3a178-2bc0-4d41-9975-aa02381236bc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 711.022686] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29103980-511a-4406-84e2-c69b742321ff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.048459] env[65522]: WARNING openstack [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 711.048459] env[65522]: WARNING openstack [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 711.060047] env[65522]: INFO nova.compute.manager [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Took 23.14 seconds to build instance. [ 711.123412] env[65522]: WARNING neutronclient.v2_0.client [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 711.124157] env[65522]: WARNING openstack [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 711.124489] env[65522]: WARNING openstack [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 711.217887] env[65522]: WARNING neutronclient.v2_0.client [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 711.218952] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 711.219379] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 711.234024] env[65522]: DEBUG nova.network.neutron [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Updating instance_info_cache with network_info: [{"id": "bb734aa2-6103-4321-b1bd-f9f5a80296ed", "address": "fa:16:3e:dc:90:05", "network": {"id": "dbe1e9ac-6466-4870-b816-a013b5c9eafa", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-2014236375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "d7f98c0ae38740df877c198e2cb611a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb734aa2-61", "ovs_interfaceid": "bb734aa2-6103-4321-b1bd-f9f5a80296ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 711.251575] env[65522]: DEBUG nova.scheduler.client.report [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 711.383414] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 711.383414] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 711.467166] env[65522]: WARNING neutronclient.v2_0.client [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 711.467677] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 711.468066] env[65522]: WARNING openstack [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 711.561039] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f427bfaa-134d-4aa9-a021-76a28d3cd94f tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "f44db343-7e17-4780-9a67-245992ac1a63" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.667s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 711.589109] env[65522]: DEBUG nova.network.neutron [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Updated VIF entry in instance network info cache for port 7d49f270-6cbd-4767-91fc-f899c46564c9. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 711.589598] env[65522]: DEBUG nova.network.neutron [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Updating instance_info_cache with network_info: [{"id": "7d49f270-6cbd-4767-91fc-f899c46564c9", "address": "fa:16:3e:af:04:9f", "network": {"id": "6a60138b-7f03-4817-ac9a-1facdc1cc0cc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1344754077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4fdbd0ed78c41a591ccc0a587313ec8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f85835c8-5d0c-4b2f-97c4-6c4006580f79", "external-id": "nsx-vlan-transportzone-245", "segmentation_id": 245, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d49f270-6c", "ovs_interfaceid": "7d49f270-6cbd-4767-91fc-f899c46564c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 711.737436] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Releasing lock "refresh_cache-c0c99008-1907-4d3a-b4f6-ee6e916e2c67" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 711.737830] env[65522]: DEBUG nova.compute.manager [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Instance network_info: |[{"id": "bb734aa2-6103-4321-b1bd-f9f5a80296ed", "address": "fa:16:3e:dc:90:05", "network": {"id": "dbe1e9ac-6466-4870-b816-a013b5c9eafa", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-2014236375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "d7f98c0ae38740df877c198e2cb611a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb734aa2-61", "ovs_interfaceid": "bb734aa2-6103-4321-b1bd-f9f5a80296ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 711.757734] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.636s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 711.758326] env[65522]: DEBUG nova.compute.manager [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 711.762452] env[65522]: DEBUG oslo_concurrency.lockutils [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.011s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 711.765128] env[65522]: INFO nova.compute.claims [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.094185] env[65522]: DEBUG oslo_concurrency.lockutils [req-11b2b4e9-413e-40ba-8530-2f0b22a9cbe8 req-1f67df5d-a4bb-483e-8fda-171d12f4208d service nova] Releasing lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 712.274729] env[65522]: DEBUG nova.compute.utils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 712.276942] env[65522]: DEBUG nova.compute.manager [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 712.277261] env[65522]: DEBUG nova.network.neutron [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 712.277687] env[65522]: WARNING neutronclient.v2_0.client [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 712.278096] env[65522]: WARNING neutronclient.v2_0.client [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 712.279031] env[65522]: WARNING openstack [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.282330] env[65522]: WARNING openstack [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.337948] env[65522]: DEBUG nova.policy [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7ef25c2fe4643748839c6fc67e74ad0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37fc01876567476f9b93d765b2cfddc8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 712.716247] env[65522]: DEBUG nova.network.neutron [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Successfully created port: 37951fc0-2513-48ca-8e6a-42be7de65465 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 712.792775] env[65522]: DEBUG nova.compute.manager [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 713.191202] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb41913-9e8e-4843-bf44-ca1b86b94099 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.202089] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c913f64-6bb9-417c-89bb-61366f37caed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.242395] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14b6888-7c41-4ed0-9f8f-8c139d072b2f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.252131] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1149ba5-a178-4449-9374-f414abf4e6b4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.271108] env[65522]: DEBUG nova.compute.provider_tree [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.775708] env[65522]: DEBUG nova.scheduler.client.report [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 713.806044] env[65522]: DEBUG nova.compute.manager [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 713.966921] env[65522]: DEBUG nova.virt.hardware [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 713.967094] env[65522]: DEBUG nova.virt.hardware [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 713.967244] env[65522]: DEBUG nova.virt.hardware [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 713.967430] env[65522]: DEBUG nova.virt.hardware [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 713.967570] env[65522]: DEBUG nova.virt.hardware [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 713.967711] env[65522]: DEBUG nova.virt.hardware [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 713.967946] env[65522]: DEBUG nova.virt.hardware [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 713.969446] env[65522]: DEBUG nova.virt.hardware [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 713.969446] env[65522]: DEBUG nova.virt.hardware [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 713.969446] env[65522]: DEBUG nova.virt.hardware [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 713.969446] env[65522]: DEBUG nova.virt.hardware [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 713.973514] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846801a5-1540-4f6a-bec4-6f41153aae14 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.987449] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5345b0-3a28-4319-adc6-9fab4e06fa8f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.996425] env[65522]: DEBUG nova.virt.hardware [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 713.997128] env[65522]: DEBUG nova.virt.hardware [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 713.997428] env[65522]: DEBUG nova.virt.hardware [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 713.999028] env[65522]: DEBUG nova.virt.hardware [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 713.999028] env[65522]: DEBUG nova.virt.hardware [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 713.999028] env[65522]: DEBUG nova.virt.hardware [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 713.999028] env[65522]: DEBUG nova.virt.hardware [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 713.999028] env[65522]: DEBUG nova.virt.hardware [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 713.999197] env[65522]: DEBUG nova.virt.hardware [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 713.999344] env[65522]: DEBUG nova.virt.hardware [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 714.001975] env[65522]: DEBUG nova.virt.hardware [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 714.002796] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54312919-4ce6-4af0-bebd-c68d15455781 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.021493] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 714.032289] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Creating folder: Project (d30f5d54634a4096801c04fc43ff7460). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 714.037127] env[65522]: DEBUG nova.virt.hardware [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 714.037592] env[65522]: DEBUG nova.virt.hardware [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 714.038046] env[65522]: DEBUG nova.virt.hardware [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 714.038467] env[65522]: DEBUG nova.virt.hardware [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 714.038757] env[65522]: DEBUG nova.virt.hardware [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 714.039076] env[65522]: DEBUG nova.virt.hardware [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 714.039353] env[65522]: DEBUG nova.virt.hardware [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 714.039542] env[65522]: DEBUG nova.virt.hardware [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 714.041751] env[65522]: DEBUG nova.virt.hardware [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 714.041751] env[65522]: DEBUG nova.virt.hardware [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 714.041751] env[65522]: DEBUG nova.virt.hardware [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 714.042026] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587a081c-5c33-47d4-9e87-760a12800ee2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.049568] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05e7d340-e507-4c3f-beff-c40d1dd548ff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.057223] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1dddde0-514b-4f07-a603-4747536ba4ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.080193] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:90:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '24376631-ee89-4ff1-b8ac-f09911fc8329', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb734aa2-6103-4321-b1bd-f9f5a80296ed', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 714.090611] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Creating folder: Project (d7f98c0ae38740df877c198e2cb611a0). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 714.090611] env[65522]: DEBUG oslo_vmware.rw_handles [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ebbd16-f407-eec0-0536-1f815ffa9ca5/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 714.090945] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8130172f-2d36-4ceb-96af-48fee5d856cc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.094135] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d124f95f-8984-439b-8a1d-940498f572d9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.099864] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a15d97-6a58-4f6a-b0fb-2fdf7d924d71 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.111030] env[65522]: DEBUG oslo_vmware.rw_handles [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ebbd16-f407-eec0-0536-1f815ffa9ca5/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 714.111030] env[65522]: ERROR oslo_vmware.rw_handles [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ebbd16-f407-eec0-0536-1f815ffa9ca5/disk-0.vmdk due to incomplete transfer. [ 714.118501] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6836355d-1279-418d-bfb1-b80c82930e91 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.125244] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Created folder: Project (d30f5d54634a4096801c04fc43ff7460) in parent group-v994660. [ 714.125244] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Creating folder: Instances. Parent ref: group-v994739. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 714.125244] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Created folder: Project (d7f98c0ae38740df877c198e2cb611a0) in parent group-v994660. [ 714.125244] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Creating folder: Instances. Parent ref: group-v994740. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 714.125244] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4d4a1213-05c7-4db2-97f0-5cef1b854b4b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.125940] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f05d982d-6dc7-4d56-8c2e-5b5f20db465d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.132472] env[65522]: DEBUG oslo_vmware.rw_handles [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ebbd16-f407-eec0-0536-1f815ffa9ca5/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 714.132472] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Uploaded image 1f04bff9-afe0-49b1-9eb7-e369b9348e76 to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 714.134133] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 714.134440] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4884b775-9082-4a99-a56a-1fff7c821485 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.139145] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Created folder: Instances in parent group-v994739. [ 714.140521] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 714.140521] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Created folder: Instances in parent group-v994740. [ 714.140521] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 714.140521] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 714.140521] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 714.140521] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-967f54b5-3eec-4b15-aabf-bdf680eb068d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.153897] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fc0b2526-9f49-45bc-908e-9f397fa54b1f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.171049] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 714.171049] env[65522]: value = "task-5113746" [ 714.171049] env[65522]: _type = "Task" [ 714.171049] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.178348] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 714.178348] env[65522]: value = "task-5113747" [ 714.178348] env[65522]: _type = "Task" [ 714.178348] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.178577] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 714.178577] env[65522]: value = "task-5113748" [ 714.178577] env[65522]: _type = "Task" [ 714.178577] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.182151] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113746, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.195515] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113748, 'name': CreateVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.199219] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113747, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.283559] env[65522]: DEBUG oslo_concurrency.lockutils [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 714.284297] env[65522]: DEBUG nova.compute.manager [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 714.433588] env[65522]: DEBUG nova.network.neutron [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Successfully updated port: 37951fc0-2513-48ca-8e6a-42be7de65465 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 714.686791] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113746, 'name': Destroy_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.705854] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113747, 'name': CreateVM_Task, 'duration_secs': 0.463158} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.711626] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 714.712467] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113748, 'name': CreateVM_Task, 'duration_secs': 0.505441} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.713500] env[65522]: DEBUG oslo_concurrency.lockutils [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.713578] env[65522]: DEBUG oslo_concurrency.lockutils [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 714.714093] env[65522]: DEBUG oslo_concurrency.lockutils [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 714.714431] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 714.714835] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ef0beba-026a-4fb0-806f-91f2b0b3f1f2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.717344] env[65522]: WARNING neutronclient.v2_0.client [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 714.717726] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.726024] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 714.726024] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5215bfa3-967c-a443-a1bd-1c96998d9771" [ 714.726024] env[65522]: _type = "Task" [ 714.726024] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.733924] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5215bfa3-967c-a443-a1bd-1c96998d9771, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.792379] env[65522]: DEBUG nova.compute.utils [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 714.793436] env[65522]: DEBUG nova.compute.manager [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 714.793688] env[65522]: DEBUG nova.network.neutron [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 714.794146] env[65522]: WARNING neutronclient.v2_0.client [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 714.794531] env[65522]: WARNING neutronclient.v2_0.client [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 714.795144] env[65522]: WARNING openstack [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 714.795494] env[65522]: WARNING openstack [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 714.861142] env[65522]: DEBUG nova.policy [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4500c1e9c924cef96d8dcbde3b0fba4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d64b94285334dd6b26928cabedbd67c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 714.938406] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "refresh_cache-75c8848b-aea9-43f1-8697-9224050d1fef" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.939031] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquired lock "refresh_cache-75c8848b-aea9-43f1-8697-9224050d1fef" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 714.939210] env[65522]: DEBUG nova.network.neutron [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 715.185265] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113746, 'name': Destroy_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.195196] env[65522]: DEBUG nova.network.neutron [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Successfully created port: d4e8efaa-adbc-4eec-adf0-e3f651352ba9 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 715.235314] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5215bfa3-967c-a443-a1bd-1c96998d9771, 'name': SearchDatastore_Task, 'duration_secs': 0.017289} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.235661] env[65522]: DEBUG oslo_concurrency.lockutils [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 715.235904] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 715.236174] env[65522]: DEBUG oslo_concurrency.lockutils [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.236323] env[65522]: DEBUG oslo_concurrency.lockutils [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 715.236552] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 715.236861] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 715.237187] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 715.237438] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e1fef6b-5248-4bf4-ab27-15bcbd444dab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.239728] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49fa7877-0590-4eec-b283-b8baf4b755e9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.248880] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Waiting for the task: (returnval){ [ 715.248880] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527b2c62-f35f-7dd8-16a9-06979eae3cf1" [ 715.248880] env[65522]: _type = "Task" [ 715.248880] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.254343] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 715.254542] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 715.258895] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d441a48-6802-4ffe-8715-a30199654230 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.261773] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527b2c62-f35f-7dd8-16a9-06979eae3cf1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.267197] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 715.267197] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5292510d-997d-5565-6fb3-20c73a184982" [ 715.267197] env[65522]: _type = "Task" [ 715.267197] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.279829] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5292510d-997d-5565-6fb3-20c73a184982, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.304610] env[65522]: DEBUG nova.compute.manager [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 715.444802] env[65522]: WARNING openstack [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 715.444802] env[65522]: WARNING openstack [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 715.687662] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113746, 'name': Destroy_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.763684] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527b2c62-f35f-7dd8-16a9-06979eae3cf1, 'name': SearchDatastore_Task, 'duration_secs': 0.034697} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.764137] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 715.764450] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 715.764719] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.780218] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5292510d-997d-5565-6fb3-20c73a184982, 'name': SearchDatastore_Task, 'duration_secs': 0.021058} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.780512] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1408fa8a-f709-496b-a35b-d7eba7c3c798 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.787325] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 715.787325] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5230c86e-7970-c80e-4dce-44187a4879fe" [ 715.787325] env[65522]: _type = "Task" [ 715.787325] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.797132] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5230c86e-7970-c80e-4dce-44187a4879fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.809147] env[65522]: DEBUG nova.network.neutron [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 715.812335] env[65522]: INFO nova.virt.block_device [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Booting with volume 78aedba8-1717-4210-afe1-44eb7ae91932 at /dev/sda [ 715.864225] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4356d460-14de-47ac-bca1-51d584e5f495 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.876181] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029b4cd9-021a-426d-89e7-50dddd5851bc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.928738] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c999aca2-d1c6-4cd2-9873-c556b07f1f99 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.939886] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea6fc8e-4ff5-4d25-91b0-b34abaf99c02 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.952716] env[65522]: WARNING openstack [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 715.952813] env[65522]: WARNING openstack [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 715.994463] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d5106b-03eb-47c7-bb2d-47320b0e2cbe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.997684] env[65522]: DEBUG oslo_vmware.rw_handles [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52037107-5b13-9e63-8be1-bcc379b5dba5/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 715.998654] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04000332-5e79-4dcb-a7a8-0c876ed4c3c9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.008044] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248c038d-bdef-4729-a3e2-ae41d5b2dec5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.010805] env[65522]: DEBUG oslo_vmware.rw_handles [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52037107-5b13-9e63-8be1-bcc379b5dba5/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 716.011022] env[65522]: ERROR oslo_vmware.rw_handles [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52037107-5b13-9e63-8be1-bcc379b5dba5/disk-0.vmdk due to incomplete transfer.: glanceclient.exc.HTTPException: HTTP 410 Gone: Image 4cd7fa95-e6d9-457c-8cf1-e3881f933468 could not be found after upload. The image may have been deleted during the upload, cleaning up the chunks uploaded. [ 716.011226] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-50cdc631-54a4-4660-9b45-51fce32639ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.023026] env[65522]: DEBUG oslo_vmware.rw_handles [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52037107-5b13-9e63-8be1-bcc379b5dba5/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 716.023026] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 716.023026] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-22604439-9e73-406f-afc7-16b3a0accf9d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.030498] env[65522]: DEBUG nova.virt.block_device [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Updating existing volume attachment record: a03e46a6-975c-46c1-b23f-1c47015f4237 {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 716.036755] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 716.036755] env[65522]: value = "task-5113749" [ 716.036755] env[65522]: _type = "Task" [ 716.036755] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.045018] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113749, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.185081] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113746, 'name': Destroy_Task, 'duration_secs': 1.616618} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.185344] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Destroyed the VM [ 716.185587] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 716.185861] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2c45f63b-9f16-4307-b406-df464a61837c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.194759] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 716.194759] env[65522]: value = "task-5113750" [ 716.194759] env[65522]: _type = "Task" [ 716.194759] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.205540] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113750, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.299218] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5230c86e-7970-c80e-4dce-44187a4879fe, 'name': SearchDatastore_Task, 'duration_secs': 0.016927} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.299560] env[65522]: DEBUG oslo_concurrency.lockutils [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 716.299877] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 6e10e846-cdb7-458e-b511-18ab1742228e/6e10e846-cdb7-458e-b511-18ab1742228e.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 716.300219] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 716.300416] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 716.300648] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a77092ec-e404-4c36-96ed-f7c381d2e3ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.303169] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e93b28d-4e3d-464b-a99c-0f71dedc54e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.312050] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 716.312050] env[65522]: value = "task-5113751" [ 716.312050] env[65522]: _type = "Task" [ 716.312050] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.317978] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 716.318204] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 716.319492] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-321f25c2-55e1-4e4f-8c83-b4720d800ade {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.325546] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113751, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.330784] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Waiting for the task: (returnval){ [ 716.330784] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b1107e-32b2-433b-8010-18585dcc563f" [ 716.330784] env[65522]: _type = "Task" [ 716.330784] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.342416] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b1107e-32b2-433b-8010-18585dcc563f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.510963] env[65522]: WARNING neutronclient.v2_0.client [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 716.511864] env[65522]: WARNING openstack [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 716.512319] env[65522]: WARNING openstack [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 716.547979] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113749, 'name': Destroy_Task, 'duration_secs': 0.419265} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.548309] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Destroyed the VM [ 716.548733] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 716.549011] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e1017f0d-e133-426e-b079-c5a1ee2dcc4c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.559942] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 716.559942] env[65522]: value = "task-5113752" [ 716.559942] env[65522]: _type = "Task" [ 716.559942] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.574536] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113752, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.711503] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113750, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.718131] env[65522]: DEBUG nova.network.neutron [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Updating instance_info_cache with network_info: [{"id": "37951fc0-2513-48ca-8e6a-42be7de65465", "address": "fa:16:3e:5e:53:01", "network": {"id": "922eeeeb-e124-4ae5-97a4-0e995b16f7ec", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-437359683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "37fc01876567476f9b93d765b2cfddc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccf76700-491b-4462-ab19-e6d3a9ff87ac", "external-id": "nsx-vlan-transportzone-956", "segmentation_id": 956, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37951fc0-25", "ovs_interfaceid": "37951fc0-2513-48ca-8e6a-42be7de65465", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 716.827714] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113751, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.845023] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b1107e-32b2-433b-8010-18585dcc563f, 'name': SearchDatastore_Task, 'duration_secs': 0.021604} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.845714] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91107a1c-fe3e-4d5a-b4da-c90c7edb4abb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.854330] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Waiting for the task: (returnval){ [ 716.854330] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5221a009-d198-06ee-193a-8552d44d744e" [ 716.854330] env[65522]: _type = "Task" [ 716.854330] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.868632] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5221a009-d198-06ee-193a-8552d44d744e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.913879] env[65522]: DEBUG nova.network.neutron [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Successfully updated port: d4e8efaa-adbc-4eec-adf0-e3f651352ba9 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 716.923982] env[65522]: DEBUG nova.compute.manager [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Received event network-changed-3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 716.924271] env[65522]: DEBUG nova.compute.manager [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Refreshing instance network info cache due to event network-changed-3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 716.924877] env[65522]: DEBUG oslo_concurrency.lockutils [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] Acquiring lock "refresh_cache-68d8d59b-4fc8-4993-8c9e-31b2329c88dd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.924877] env[65522]: DEBUG oslo_concurrency.lockutils [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] Acquired lock "refresh_cache-68d8d59b-4fc8-4993-8c9e-31b2329c88dd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 716.925090] env[65522]: DEBUG nova.network.neutron [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Refreshing network info cache for port 3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 717.074948] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113752, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.101778] env[65522]: DEBUG nova.compute.manager [req-8d5d075d-ec40-4fe1-b532-34eda9deac7b req-0a1876b4-8d0d-4388-9979-b3d7b6352078 service nova] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Received event network-vif-plugged-bb734aa2-6103-4321-b1bd-f9f5a80296ed {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 717.102042] env[65522]: DEBUG oslo_concurrency.lockutils [req-8d5d075d-ec40-4fe1-b532-34eda9deac7b req-0a1876b4-8d0d-4388-9979-b3d7b6352078 service nova] Acquiring lock "c0c99008-1907-4d3a-b4f6-ee6e916e2c67-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 717.102280] env[65522]: DEBUG oslo_concurrency.lockutils [req-8d5d075d-ec40-4fe1-b532-34eda9deac7b req-0a1876b4-8d0d-4388-9979-b3d7b6352078 service nova] Lock "c0c99008-1907-4d3a-b4f6-ee6e916e2c67-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 717.102464] env[65522]: DEBUG oslo_concurrency.lockutils [req-8d5d075d-ec40-4fe1-b532-34eda9deac7b req-0a1876b4-8d0d-4388-9979-b3d7b6352078 service nova] Lock "c0c99008-1907-4d3a-b4f6-ee6e916e2c67-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 717.102662] env[65522]: DEBUG nova.compute.manager [req-8d5d075d-ec40-4fe1-b532-34eda9deac7b req-0a1876b4-8d0d-4388-9979-b3d7b6352078 service nova] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] No waiting events found dispatching network-vif-plugged-bb734aa2-6103-4321-b1bd-f9f5a80296ed {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 717.102859] env[65522]: WARNING nova.compute.manager [req-8d5d075d-ec40-4fe1-b532-34eda9deac7b req-0a1876b4-8d0d-4388-9979-b3d7b6352078 service nova] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Received unexpected event network-vif-plugged-bb734aa2-6103-4321-b1bd-f9f5a80296ed for instance with vm_state building and task_state spawning. [ 717.208556] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113750, 'name': RemoveSnapshot_Task} progress is 70%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.222893] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Releasing lock "refresh_cache-75c8848b-aea9-43f1-8697-9224050d1fef" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 717.224630] env[65522]: DEBUG nova.compute.manager [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Instance network_info: |[{"id": "37951fc0-2513-48ca-8e6a-42be7de65465", "address": "fa:16:3e:5e:53:01", "network": {"id": "922eeeeb-e124-4ae5-97a4-0e995b16f7ec", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-437359683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "37fc01876567476f9b93d765b2cfddc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccf76700-491b-4462-ab19-e6d3a9ff87ac", "external-id": "nsx-vlan-transportzone-956", "segmentation_id": 956, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37951fc0-25", "ovs_interfaceid": "37951fc0-2513-48ca-8e6a-42be7de65465", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 717.224952] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:53:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccf76700-491b-4462-ab19-e6d3a9ff87ac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37951fc0-2513-48ca-8e6a-42be7de65465', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 717.234670] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Creating folder: Project (37fc01876567476f9b93d765b2cfddc8). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 717.237505] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ab85bbd3-c8fe-49ce-a8f6-b73773508a72 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.249869] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Created folder: Project (37fc01876567476f9b93d765b2cfddc8) in parent group-v994660. [ 717.250203] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Creating folder: Instances. Parent ref: group-v994745. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 717.250686] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ae72a1d5-e150-4b29-ad09-74dfe23f7bf5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.265058] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Created folder: Instances in parent group-v994745. [ 717.265345] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 717.265604] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 717.265826] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-30dc0042-a244-49b7-9e6b-658cd9253d54 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.289085] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 717.289085] env[65522]: value = "task-5113755" [ 717.289085] env[65522]: _type = "Task" [ 717.289085] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.299145] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113755, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.324720] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113751, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.591829} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.325113] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 6e10e846-cdb7-458e-b511-18ab1742228e/6e10e846-cdb7-458e-b511-18ab1742228e.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 717.325267] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 717.325537] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d87700c-4135-498d-9a24-cab5e3587664 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.334937] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 717.334937] env[65522]: value = "task-5113756" [ 717.334937] env[65522]: _type = "Task" [ 717.334937] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.346868] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113756, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.366510] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5221a009-d198-06ee-193a-8552d44d744e, 'name': SearchDatastore_Task, 'duration_secs': 0.033832} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.366808] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 717.367090] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] c0c99008-1907-4d3a-b4f6-ee6e916e2c67/c0c99008-1907-4d3a-b4f6-ee6e916e2c67.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 717.367380] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-93d44143-e3ae-45c9-a686-082c931e2281 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.377223] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Waiting for the task: (returnval){ [ 717.377223] env[65522]: value = "task-5113757" [ 717.377223] env[65522]: _type = "Task" [ 717.377223] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.387993] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113757, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.417080] env[65522]: DEBUG oslo_concurrency.lockutils [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Acquiring lock "refresh_cache-42a08cfb-d865-4967-a086-370a2ca98b7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.417407] env[65522]: DEBUG oslo_concurrency.lockutils [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Acquired lock "refresh_cache-42a08cfb-d865-4967-a086-370a2ca98b7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 717.417644] env[65522]: DEBUG nova.network.neutron [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 717.429106] env[65522]: WARNING neutronclient.v2_0.client [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 717.429562] env[65522]: WARNING openstack [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.429946] env[65522]: WARNING openstack [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.575207] env[65522]: DEBUG oslo_vmware.api [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113752, 'name': RemoveSnapshot_Task, 'duration_secs': 0.952233} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.575528] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 717.575757] env[65522]: DEBUG nova.compute.utils [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Cleaning up image 4cd7fa95-e6d9-457c-8cf1-e3881f933468 {{(pid=65522) delete_image /opt/stack/nova/nova/compute/utils.py:1352}} [ 717.625022] env[65522]: WARNING openstack [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.625022] env[65522]: WARNING openstack [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.716118] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113750, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.769599] env[65522]: WARNING neutronclient.v2_0.client [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 717.770396] env[65522]: WARNING openstack [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.770773] env[65522]: WARNING openstack [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.802039] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113755, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.853102] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113756, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.218422} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.853264] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 717.854233] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6841e55-fe5b-4397-a022-7c686fba26da {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.885726] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] 6e10e846-cdb7-458e-b511-18ab1742228e/6e10e846-cdb7-458e-b511-18ab1742228e.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 717.886220] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50238c72-0d95-442f-9b50-1da5f47eae98 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.915682] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113757, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.917953] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 717.917953] env[65522]: value = "task-5113758" [ 717.917953] env[65522]: _type = "Task" [ 717.917953] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.923181] env[65522]: WARNING openstack [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 717.923792] env[65522]: WARNING openstack [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 717.941115] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Acquiring lock "20855931-a02d-4b53-b998-3d89d4ff5d1a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 717.941480] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Lock "20855931-a02d-4b53-b998-3d89d4ff5d1a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 717.962517] env[65522]: DEBUG nova.network.neutron [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Updated VIF entry in instance network info cache for port 3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 717.962890] env[65522]: DEBUG nova.network.neutron [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Updating instance_info_cache with network_info: [{"id": "3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6", "address": "fa:16:3e:6f:2f:ba", "network": {"id": "bed09e85-fdc4-4258-9be8-027aa4cf34a6", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-752246656-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb9a9e5da1194ab7927f72d21d1484b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de5fcb06-b0d0-467f-86fe-06882165ac31", "external-id": "nsx-vlan-transportzone-595", "segmentation_id": 595, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bbf4b9c-9f", "ovs_interfaceid": "3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 717.995163] env[65522]: DEBUG nova.network.neutron [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 718.024342] env[65522]: WARNING openstack [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 718.024744] env[65522]: WARNING openstack [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 718.105466] env[65522]: WARNING neutronclient.v2_0.client [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 718.106291] env[65522]: WARNING openstack [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 718.106680] env[65522]: WARNING openstack [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 718.168616] env[65522]: DEBUG nova.compute.manager [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 718.169348] env[65522]: DEBUG nova.virt.hardware [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 718.169348] env[65522]: DEBUG nova.virt.hardware [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 718.169568] env[65522]: DEBUG nova.virt.hardware [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 718.169594] env[65522]: DEBUG nova.virt.hardware [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 718.173016] env[65522]: DEBUG nova.virt.hardware [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 718.173016] env[65522]: DEBUG nova.virt.hardware [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 718.173016] env[65522]: DEBUG nova.virt.hardware [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.173016] env[65522]: DEBUG nova.virt.hardware [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 718.173016] env[65522]: DEBUG nova.virt.hardware [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 718.173495] env[65522]: DEBUG nova.virt.hardware [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 718.173495] env[65522]: DEBUG nova.virt.hardware [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 718.173495] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-152b32a0-043f-4e6f-9492-7e221f63dd8f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.195827] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "361e54c7-4309-48b3-ad3e-b2ebc783741c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.195827] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "361e54c7-4309-48b3-ad3e-b2ebc783741c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.200226] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae65e6c-6e81-479d-bdf7-f0ac949a59fd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.224497] env[65522]: DEBUG oslo_vmware.api [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113750, 'name': RemoveSnapshot_Task, 'duration_secs': 1.54707} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.224800] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 718.225044] env[65522]: INFO nova.compute.manager [None req-09981380-10e7-47cf-acb5-0a229a3c8d22 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Took 19.25 seconds to snapshot the instance on the hypervisor. [ 718.234748] env[65522]: DEBUG nova.network.neutron [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Updating instance_info_cache with network_info: [{"id": "d4e8efaa-adbc-4eec-adf0-e3f651352ba9", "address": "fa:16:3e:b9:0a:a7", "network": {"id": "2bb5d3a6-95f2-43ba-9606-ad864f78a377", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-2069424406-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d64b94285334dd6b26928cabedbd67c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4e8efaa-ad", "ovs_interfaceid": "d4e8efaa-adbc-4eec-adf0-e3f651352ba9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 718.301745] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113755, 'name': CreateVM_Task, 'duration_secs': 0.580827} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.301881] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 718.302422] env[65522]: WARNING neutronclient.v2_0.client [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 718.302806] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.302975] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 718.303303] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 718.303577] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a0b659b-9fc3-4c4c-8037-7a7c0d5a86fd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.313064] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 718.313064] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5266aa6f-1c92-eebd-aa0a-638117b849bc" [ 718.313064] env[65522]: _type = "Task" [ 718.313064] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.324426] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5266aa6f-1c92-eebd-aa0a-638117b849bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.397883] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113757, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.697504} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.398264] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] c0c99008-1907-4d3a-b4f6-ee6e916e2c67/c0c99008-1907-4d3a-b4f6-ee6e916e2c67.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 718.398432] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 718.398761] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea8d5a27-1354-4801-ae97-4e9c1f92e1a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.407980] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Waiting for the task: (returnval){ [ 718.407980] env[65522]: value = "task-5113759" [ 718.407980] env[65522]: _type = "Task" [ 718.407980] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.419137] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113759, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.429719] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113758, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.445394] env[65522]: DEBUG nova.compute.manager [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 718.465132] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "428873b8-1e58-4924-86f3-8feca8f53342" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.465379] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "428873b8-1e58-4924-86f3-8feca8f53342" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.466767] env[65522]: DEBUG oslo_concurrency.lockutils [req-97f4b577-a165-46f9-9ea0-013c1e263d67 req-ab9bcb1d-0336-436e-98c3-409b777e2753 service nova] Releasing lock "refresh_cache-68d8d59b-4fc8-4993-8c9e-31b2329c88dd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 718.540920] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.541149] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.595858] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "028c7fc5-5624-4dd5-9e2f-48191a86f765" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.597422] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "028c7fc5-5624-4dd5-9e2f-48191a86f765" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.691945] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0bab39-82da-479d-aaaa-dec98abbe69b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.701234] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a11d0c19-e3e4-4b0e-8ff3-01c337e5776a tempest-ServersAdminNegativeTestJSON-613751407 tempest-ServersAdminNegativeTestJSON-613751407-project-admin] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Suspending the VM {{(pid=65522) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 718.701586] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-54110e24-504a-4bc9-9834-760ea755f076 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.705893] env[65522]: DEBUG nova.compute.manager [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 718.718724] env[65522]: DEBUG oslo_vmware.api [None req-a11d0c19-e3e4-4b0e-8ff3-01c337e5776a tempest-ServersAdminNegativeTestJSON-613751407 tempest-ServersAdminNegativeTestJSON-613751407-project-admin] Waiting for the task: (returnval){ [ 718.718724] env[65522]: value = "task-5113760" [ 718.718724] env[65522]: _type = "Task" [ 718.718724] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.735901] env[65522]: DEBUG oslo_vmware.api [None req-a11d0c19-e3e4-4b0e-8ff3-01c337e5776a tempest-ServersAdminNegativeTestJSON-613751407 tempest-ServersAdminNegativeTestJSON-613751407-project-admin] Task: {'id': task-5113760, 'name': SuspendVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.738422] env[65522]: DEBUG oslo_concurrency.lockutils [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Releasing lock "refresh_cache-42a08cfb-d865-4967-a086-370a2ca98b7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 718.743182] env[65522]: DEBUG nova.compute.manager [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Instance network_info: |[{"id": "d4e8efaa-adbc-4eec-adf0-e3f651352ba9", "address": "fa:16:3e:b9:0a:a7", "network": {"id": "2bb5d3a6-95f2-43ba-9606-ad864f78a377", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-2069424406-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d64b94285334dd6b26928cabedbd67c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4e8efaa-ad", "ovs_interfaceid": "d4e8efaa-adbc-4eec-adf0-e3f651352ba9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 718.743374] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:0a:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48512b02-ad5c-4105-ba7d-fd4775acf8e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd4e8efaa-adbc-4eec-adf0-e3f651352ba9', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 718.748986] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Creating folder: Project (6d64b94285334dd6b26928cabedbd67c). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 718.749352] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2553b8ef-e094-4fb3-836d-60b6aad03917 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.768039] env[65522]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 718.768588] env[65522]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=65522) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 718.768827] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Folder already exists: Project (6d64b94285334dd6b26928cabedbd67c). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 718.769200] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Creating folder: Instances. Parent ref: group-v994725. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 718.769425] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-226c6a72-a276-49f7-bb61-69a08b5f3afc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.785035] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Created folder: Instances in parent group-v994725. [ 718.786574] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 718.786574] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 718.787782] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-750832f1-f6c3-496b-8b31-32173b29da8b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.814648] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 718.814648] env[65522]: value = "task-5113763" [ 718.814648] env[65522]: _type = "Task" [ 718.814648] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.830254] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5266aa6f-1c92-eebd-aa0a-638117b849bc, 'name': SearchDatastore_Task, 'duration_secs': 0.018992} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.832859] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 718.833681] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 718.833681] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.833681] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 718.833826] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 718.834120] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113763, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.834320] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec46ea2c-cfd3-4e4b-90ab-a16b3bad34f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.845675] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 718.846041] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 718.847083] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65bdc5cd-61f7-495f-ab49-2f6966599b8e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.859201] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 718.859201] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f9d296-c0fe-f13b-102d-572bf4ec2a32" [ 718.859201] env[65522]: _type = "Task" [ 718.859201] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.870450] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f9d296-c0fe-f13b-102d-572bf4ec2a32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.920922] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113759, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.371577} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.925380] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 718.925549] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8bcb35b-de60-4593-837e-9a939f7b01c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.936414] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113758, 'name': ReconfigVM_Task, 'duration_secs': 0.703925} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.946847] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Reconfigured VM instance instance-0000001c to attach disk [datastore1] 6e10e846-cdb7-458e-b511-18ab1742228e/6e10e846-cdb7-458e-b511-18ab1742228e.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 718.957727] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Reconfiguring VM instance instance-0000001d to attach disk [datastore1] c0c99008-1907-4d3a-b4f6-ee6e916e2c67/c0c99008-1907-4d3a-b4f6-ee6e916e2c67.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 718.958069] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4b9b199f-fed4-4f4d-9beb-968d02ec3f60 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.963172] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32ffb2a7-4202-4af2-9248-7ad4f343f726 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.982553] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 718.992193] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Waiting for the task: (returnval){ [ 718.992193] env[65522]: value = "task-5113765" [ 718.992193] env[65522]: _type = "Task" [ 718.992193] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.996042] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 718.996042] env[65522]: value = "task-5113764" [ 718.996042] env[65522]: _type = "Task" [ 718.996042] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.011209] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113764, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.013090] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 719.013387] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 719.015967] env[65522]: INFO nova.compute.claims [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 719.023590] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113765, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.047039] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 719.101378] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 719.121490] env[65522]: DEBUG oslo_concurrency.lockutils [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 719.230479] env[65522]: DEBUG oslo_vmware.api [None req-a11d0c19-e3e4-4b0e-8ff3-01c337e5776a tempest-ServersAdminNegativeTestJSON-613751407 tempest-ServersAdminNegativeTestJSON-613751407-project-admin] Task: {'id': task-5113760, 'name': SuspendVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.234070] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 719.333428] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113763, 'name': CreateVM_Task, 'duration_secs': 0.502161} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.333428] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 719.333428] env[65522]: WARNING neutronclient.v2_0.client [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 719.333604] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': True, 'boot_index': 0, 'mount_device': '/dev/sda', 'disk_bus': None, 'guest_format': None, 'device_type': None, 'attachment_id': 'a03e46a6-975c-46c1-b23f-1c47015f4237', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994729', 'volume_id': '78aedba8-1717-4210-afe1-44eb7ae91932', 'name': 'volume-78aedba8-1717-4210-afe1-44eb7ae91932', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '42a08cfb-d865-4967-a086-370a2ca98b7d', 'attached_at': '', 'detached_at': '', 'volume_id': '78aedba8-1717-4210-afe1-44eb7ae91932', 'serial': '78aedba8-1717-4210-afe1-44eb7ae91932'}, 'volume_type': None}], 'swap': None} {{(pid=65522) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 719.333856] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Root volume attach. Driver type: vmdk {{(pid=65522) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 719.334866] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bfc98d-1be3-4edb-8b88-80220b169f55 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.346751] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2bcb9cc-c1d1-4f28-8ee1-05fd4d8acef5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.356739] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1abbb5-86bf-445f-af8e-42ee8efd4d11 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.374844] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-089447b3-7a3e-467e-b4fd-7af77d0a3ee4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.377712] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f9d296-c0fe-f13b-102d-572bf4ec2a32, 'name': SearchDatastore_Task, 'duration_secs': 0.01764} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.379439] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5755884-6a12-4e91-bc6a-fb0857a9ea58 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.385193] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Waiting for the task: (returnval){ [ 719.385193] env[65522]: value = "task-5113766" [ 719.385193] env[65522]: _type = "Task" [ 719.385193] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.386768] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 719.386768] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52990170-721d-8b87-c5d1-ea9e062c71e1" [ 719.386768] env[65522]: _type = "Task" [ 719.386768] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.402245] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113766, 'name': RelocateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.405940] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52990170-721d-8b87-c5d1-ea9e062c71e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.507223] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 719.511556] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113765, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.514869] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113764, 'name': Rename_Task, 'duration_secs': 0.288811} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.515237] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 719.515470] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67710682-7d94-4baf-b0d9-151f8698cca3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.523517] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 719.523517] env[65522]: value = "task-5113767" [ 719.523517] env[65522]: _type = "Task" [ 719.523517] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.537532] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113767, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.569495] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 719.625887] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 719.733364] env[65522]: DEBUG oslo_vmware.api [None req-a11d0c19-e3e4-4b0e-8ff3-01c337e5776a tempest-ServersAdminNegativeTestJSON-613751407 tempest-ServersAdminNegativeTestJSON-613751407-project-admin] Task: {'id': task-5113760, 'name': SuspendVM_Task, 'duration_secs': 0.940324} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.733873] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a11d0c19-e3e4-4b0e-8ff3-01c337e5776a tempest-ServersAdminNegativeTestJSON-613751407 tempest-ServersAdminNegativeTestJSON-613751407-project-admin] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Suspended the VM {{(pid=65522) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 719.734255] env[65522]: DEBUG nova.compute.manager [None req-a11d0c19-e3e4-4b0e-8ff3-01c337e5776a tempest-ServersAdminNegativeTestJSON-613751407 tempest-ServersAdminNegativeTestJSON-613751407-project-admin] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 719.735363] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dcb9119-94cf-4bf1-a09f-6e30a5e992fa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.830414] env[65522]: DEBUG nova.compute.manager [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Stashing vm_state: active {{(pid=65522) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 719.839686] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "da211708-a6a2-4e03-bb41-c5b93564de20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 719.839996] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "da211708-a6a2-4e03-bb41-c5b93564de20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 719.907268] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113766, 'name': RelocateVM_Task} progress is 38%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.910739] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52990170-721d-8b87-c5d1-ea9e062c71e1, 'name': SearchDatastore_Task, 'duration_secs': 0.01909} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.910739] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 719.910739] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 75c8848b-aea9-43f1-8697-9224050d1fef/75c8848b-aea9-43f1-8697-9224050d1fef.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 719.911114] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4bdfcc0-5520-414a-8e90-90890a6d7d29 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.922194] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 719.922194] env[65522]: value = "task-5113768" [ 719.922194] env[65522]: _type = "Task" [ 719.922194] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.932517] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5113768, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.006962] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113765, 'name': ReconfigVM_Task, 'duration_secs': 0.645286} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.007333] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Reconfigured VM instance instance-0000001d to attach disk [datastore1] c0c99008-1907-4d3a-b4f6-ee6e916e2c67/c0c99008-1907-4d3a-b4f6-ee6e916e2c67.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 720.008108] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5340851-dd8f-495f-98b3-3238af979cab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.017922] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Waiting for the task: (returnval){ [ 720.017922] env[65522]: value = "task-5113769" [ 720.017922] env[65522]: _type = "Task" [ 720.017922] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.034578] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113769, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.042491] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113767, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.095024] env[65522]: DEBUG nova.compute.manager [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Received event network-changed-7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 720.095205] env[65522]: DEBUG nova.compute.manager [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Refreshing instance network info cache due to event network-changed-7d49f270-6cbd-4767-91fc-f899c46564c9. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 720.095324] env[65522]: DEBUG oslo_concurrency.lockutils [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] Acquiring lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.095471] env[65522]: DEBUG oslo_concurrency.lockutils [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] Acquired lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 720.095693] env[65522]: DEBUG nova.network.neutron [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Refreshing network info cache for port 7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 720.183756] env[65522]: DEBUG nova.compute.manager [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Received event network-changed-bb734aa2-6103-4321-b1bd-f9f5a80296ed {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 720.183756] env[65522]: DEBUG nova.compute.manager [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Refreshing instance network info cache due to event network-changed-bb734aa2-6103-4321-b1bd-f9f5a80296ed. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 720.184011] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Acquiring lock "refresh_cache-c0c99008-1907-4d3a-b4f6-ee6e916e2c67" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.184137] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Acquired lock "refresh_cache-c0c99008-1907-4d3a-b4f6-ee6e916e2c67" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 720.184295] env[65522]: DEBUG nova.network.neutron [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Refreshing network info cache for port bb734aa2-6103-4321-b1bd-f9f5a80296ed {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 720.345524] env[65522]: DEBUG nova.compute.manager [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 720.357417] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 720.416423] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113766, 'name': RelocateVM_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.439137] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5113768, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.538138] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113769, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.549618] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113767, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.603113] env[65522]: WARNING neutronclient.v2_0.client [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 720.603847] env[65522]: WARNING openstack [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.604355] env[65522]: WARNING openstack [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.613731] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c23ec4a-1265-4be5-99bc-4a7e32ffe905 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.627034] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec62898b-e5db-4a18-9f8b-3468ed6a31cd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.666626] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c74bc157-ddb3-4ae2-805b-ed5c02d6b6bd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.677186] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfbee213-7f47-45e5-b15d-7bcaa1bcfa4d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.694737] env[65522]: WARNING neutronclient.v2_0.client [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 720.695532] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.695911] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.706605] env[65522]: DEBUG nova.compute.provider_tree [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.850320] env[65522]: WARNING openstack [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.850854] env[65522]: WARNING openstack [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.864469] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.865648] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.896387] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 720.904029] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113766, 'name': RelocateVM_Task} progress is 63%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.938525] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5113768, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.955859] env[65522]: WARNING neutronclient.v2_0.client [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 720.956718] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.957220] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.979708] env[65522]: WARNING neutronclient.v2_0.client [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 720.980381] env[65522]: WARNING openstack [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.980710] env[65522]: WARNING openstack [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.035728] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113769, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.051345] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113767, 'name': PowerOnVM_Task} progress is 71%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.087204] env[65522]: DEBUG nova.network.neutron [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Updated VIF entry in instance network info cache for port bb734aa2-6103-4321-b1bd-f9f5a80296ed. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 721.087657] env[65522]: DEBUG nova.network.neutron [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Updating instance_info_cache with network_info: [{"id": "bb734aa2-6103-4321-b1bd-f9f5a80296ed", "address": "fa:16:3e:dc:90:05", "network": {"id": "dbe1e9ac-6466-4870-b816-a013b5c9eafa", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-2014236375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d7f98c0ae38740df877c198e2cb611a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "24376631-ee89-4ff1-b8ac-f09911fc8329", "external-id": "nsx-vlan-transportzone-960", "segmentation_id": 960, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb734aa2-61", "ovs_interfaceid": "bb734aa2-6103-4321-b1bd-f9f5a80296ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 721.092483] env[65522]: DEBUG nova.network.neutron [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Updated VIF entry in instance network info cache for port 7d49f270-6cbd-4767-91fc-f899c46564c9. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 721.092864] env[65522]: DEBUG nova.network.neutron [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Updating instance_info_cache with network_info: [{"id": "7d49f270-6cbd-4767-91fc-f899c46564c9", "address": "fa:16:3e:af:04:9f", "network": {"id": "6a60138b-7f03-4817-ac9a-1facdc1cc0cc", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1344754077-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4fdbd0ed78c41a591ccc0a587313ec8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f85835c8-5d0c-4b2f-97c4-6c4006580f79", "external-id": "nsx-vlan-transportzone-245", "segmentation_id": 245, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d49f270-6c", "ovs_interfaceid": "7d49f270-6cbd-4767-91fc-f899c46564c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 721.216040] env[65522]: DEBUG nova.scheduler.client.report [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 721.383548] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquiring lock "dbf29bd1-780c-4756-a203-4fad05a9d3b9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 721.383859] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lock "dbf29bd1-780c-4756-a203-4fad05a9d3b9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.384121] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquiring lock "dbf29bd1-780c-4756-a203-4fad05a9d3b9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 721.384422] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lock "dbf29bd1-780c-4756-a203-4fad05a9d3b9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.384678] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lock "dbf29bd1-780c-4756-a203-4fad05a9d3b9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.388666] env[65522]: INFO nova.compute.manager [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Terminating instance [ 721.407586] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113766, 'name': RelocateVM_Task} progress is 76%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.437687] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5113768, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.472170] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "6c8165f5-6769-4954-b9ac-de8551954377" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 721.473081] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "6c8165f5-6769-4954-b9ac-de8551954377" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.473081] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "6c8165f5-6769-4954-b9ac-de8551954377-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 721.473081] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "6c8165f5-6769-4954-b9ac-de8551954377-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.473325] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "6c8165f5-6769-4954-b9ac-de8551954377-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.477459] env[65522]: INFO nova.compute.manager [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Terminating instance [ 721.531256] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113769, 'name': Rename_Task, 'duration_secs': 1.204046} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.535236] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 721.535632] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8957f280-d731-4bf0-b5f6-c15fcaae411c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.546800] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113767, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.548742] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Waiting for the task: (returnval){ [ 721.548742] env[65522]: value = "task-5113770" [ 721.548742] env[65522]: _type = "Task" [ 721.548742] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.558850] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113770, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.591990] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Releasing lock "refresh_cache-c0c99008-1907-4d3a-b4f6-ee6e916e2c67" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 721.592623] env[65522]: DEBUG nova.compute.manager [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Received event network-vif-plugged-37951fc0-2513-48ca-8e6a-42be7de65465 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 721.592623] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Acquiring lock "75c8848b-aea9-43f1-8697-9224050d1fef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 721.592860] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Lock "75c8848b-aea9-43f1-8697-9224050d1fef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.593088] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Lock "75c8848b-aea9-43f1-8697-9224050d1fef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.593278] env[65522]: DEBUG nova.compute.manager [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] No waiting events found dispatching network-vif-plugged-37951fc0-2513-48ca-8e6a-42be7de65465 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 721.593443] env[65522]: WARNING nova.compute.manager [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Received unexpected event network-vif-plugged-37951fc0-2513-48ca-8e6a-42be7de65465 for instance with vm_state building and task_state spawning. [ 721.594103] env[65522]: DEBUG nova.compute.manager [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Received event network-changed-37951fc0-2513-48ca-8e6a-42be7de65465 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 721.594103] env[65522]: DEBUG nova.compute.manager [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Refreshing instance network info cache due to event network-changed-37951fc0-2513-48ca-8e6a-42be7de65465. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 721.594103] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Acquiring lock "refresh_cache-75c8848b-aea9-43f1-8697-9224050d1fef" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.594103] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Acquired lock "refresh_cache-75c8848b-aea9-43f1-8697-9224050d1fef" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.594350] env[65522]: DEBUG nova.network.neutron [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Refreshing network info cache for port 37951fc0-2513-48ca-8e6a-42be7de65465 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 721.599903] env[65522]: DEBUG oslo_concurrency.lockutils [req-a5f53313-88bd-477e-8cc6-987b03ae4f27 req-29e99e6d-a0b5-41cd-94b4-605db4e7ec96 service nova] Releasing lock "refresh_cache-57d32fb8-885e-44ba-967b-69f4bd1f5744" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 721.723988] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.710s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.728403] env[65522]: DEBUG nova.compute.manager [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 721.732146] env[65522]: DEBUG oslo_concurrency.lockutils [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.611s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.882287] env[65522]: INFO nova.compute.manager [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Rescuing [ 721.882575] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "refresh_cache-ca964440-5375-4aff-8b45-96fbe829dd16" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.882744] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "refresh_cache-ca964440-5375-4aff-8b45-96fbe829dd16" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.882905] env[65522]: DEBUG nova.network.neutron [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 721.900571] env[65522]: DEBUG nova.compute.manager [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 721.901854] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 721.903523] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0aec97-5e6a-4ff7-8ca9-2b60c4fb2f9b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.912454] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113766, 'name': RelocateVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.918989] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 721.923359] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6923dbd7-817c-4e71-b252-fec449c76192 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.934478] env[65522]: DEBUG oslo_vmware.api [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 721.934478] env[65522]: value = "task-5113771" [ 721.934478] env[65522]: _type = "Task" [ 721.934478] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.945362] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5113768, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.679377} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.949962] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 75c8848b-aea9-43f1-8697-9224050d1fef/75c8848b-aea9-43f1-8697-9224050d1fef.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 721.949962] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 721.950838] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-452ff2a2-7fe7-4564-bf1a-1850f1f9b85d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.956816] env[65522]: DEBUG oslo_vmware.api [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113771, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.964234] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 721.964234] env[65522]: value = "task-5113772" [ 721.964234] env[65522]: _type = "Task" [ 721.964234] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.978708] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5113772, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.985052] env[65522]: DEBUG nova.compute.manager [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 721.985667] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 721.990571] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5898c821-3a9a-432a-98b5-4ac3037efddd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.006187] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 722.006928] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-437115e3-7f1a-49b9-a0e2-5d5d7edd529e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.023052] env[65522]: DEBUG oslo_vmware.api [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 722.023052] env[65522]: value = "task-5113773" [ 722.023052] env[65522]: _type = "Task" [ 722.023052] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.042284] env[65522]: DEBUG oslo_vmware.api [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113773, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.046705] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113767, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.062122] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113770, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.097991] env[65522]: WARNING neutronclient.v2_0.client [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 722.099121] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.099793] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 722.246228] env[65522]: DEBUG nova.compute.utils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 722.251265] env[65522]: DEBUG nova.compute.manager [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 722.251265] env[65522]: DEBUG nova.network.neutron [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 722.251265] env[65522]: WARNING neutronclient.v2_0.client [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 722.251265] env[65522]: WARNING neutronclient.v2_0.client [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 722.251707] env[65522]: WARNING openstack [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.252120] env[65522]: WARNING openstack [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 722.374726] env[65522]: DEBUG nova.policy [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d0dbcb1e90647448f7e7b4a3a27811b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd5d40deeddf430baae1517fa72e2214', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 722.386703] env[65522]: WARNING neutronclient.v2_0.client [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 722.387433] env[65522]: WARNING openstack [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.387802] env[65522]: WARNING openstack [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 722.396216] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.396609] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 722.404844] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b7dc28-1254-4dfd-9e92-d54053f432ff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.423382] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113766, 'name': RelocateVM_Task} progress is 97%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.424603] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b379826-fbf5-4d99-9a24-a8315965dcb3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.461377] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1365feb9-b7e5-4243-8896-4c1df26b42c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.474711] env[65522]: DEBUG oslo_vmware.api [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113771, 'name': PowerOffVM_Task, 'duration_secs': 0.339368} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.475556] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 722.475747] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 722.477198] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ebacb1e-be0a-4e40-8572-d035281838d5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.484362] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2eca7ec8-2f7d-43bd-b9b1-874fa510208c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.486031] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5113772, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.156125} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.486670] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 722.487927] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e669bf-8896-4e19-bf27-5dc41e50f1ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.500097] env[65522]: DEBUG nova.compute.provider_tree [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.501840] env[65522]: INFO nova.compute.manager [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Rebuilding instance [ 722.522861] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Reconfiguring VM instance instance-0000001e to attach disk [datastore1] 75c8848b-aea9-43f1-8697-9224050d1fef/75c8848b-aea9-43f1-8697-9224050d1fef.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 722.523934] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18a0148b-0f6a-42d9-9416-c6b2494d362e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.560696] env[65522]: DEBUG oslo_vmware.api [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113773, 'name': PowerOffVM_Task, 'duration_secs': 0.506209} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.561059] env[65522]: DEBUG oslo_vmware.api [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113767, 'name': PowerOnVM_Task, 'duration_secs': 2.691494} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.566652] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 722.566838] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 722.567118] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 722.567315] env[65522]: INFO nova.compute.manager [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Took 16.45 seconds to spawn the instance on the hypervisor. [ 722.567536] env[65522]: DEBUG nova.compute.manager [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 722.568290] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 722.568290] env[65522]: value = "task-5113775" [ 722.568290] env[65522]: _type = "Task" [ 722.568290] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.568762] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f712391b-4bdc-4b0b-8af8-83c8786900cb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.571353] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed30b88f-7adc-412f-94cf-88b153287ef4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.581299] env[65522]: DEBUG oslo_vmware.api [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113770, 'name': PowerOnVM_Task, 'duration_secs': 0.624953} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.583860] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 722.584229] env[65522]: INFO nova.compute.manager [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Took 13.60 seconds to spawn the instance on the hypervisor. [ 722.584497] env[65522]: DEBUG nova.compute.manager [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 722.584953] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 722.585270] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 722.585546] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Deleting the datastore file [datastore1] dbf29bd1-780c-4756-a203-4fad05a9d3b9 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 722.587210] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3aac5cf-61cd-4a15-8565-66267cec5a98 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.590592] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-43816175-edb9-4020-9fa1-98d428ffd50a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.600521] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5113775, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.611104] env[65522]: DEBUG nova.compute.manager [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 722.612264] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3db4af-9bb4-4457-ab46-08dc51f49c75 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.619265] env[65522]: DEBUG oslo_vmware.api [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for the task: (returnval){ [ 722.619265] env[65522]: value = "task-5113777" [ 722.619265] env[65522]: _type = "Task" [ 722.619265] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.637971] env[65522]: DEBUG oslo_vmware.api [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113777, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.658090] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 722.658399] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 722.658627] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Deleting the datastore file [datastore2] 6c8165f5-6769-4954-b9ac-de8551954377 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 722.659780] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3bf5fb49-ba89-4936-a9ce-8774e9bbc629 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.669123] env[65522]: DEBUG oslo_vmware.api [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 722.669123] env[65522]: value = "task-5113778" [ 722.669123] env[65522]: _type = "Task" [ 722.669123] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.687235] env[65522]: DEBUG oslo_vmware.api [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113778, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.778135] env[65522]: DEBUG nova.compute.manager [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 722.884779] env[65522]: WARNING neutronclient.v2_0.client [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 722.885827] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.886461] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 722.900048] env[65522]: DEBUG nova.network.neutron [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Successfully created port: b7e73ab8-d1b6-4e32-845a-fb308037ddf9 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 722.913610] env[65522]: DEBUG nova.objects.instance [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Lazy-loading 'flavor' on Instance uuid 6abf58ca-ad6b-4fe8-844a-ca33e62a844d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 722.920813] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113766, 'name': RelocateVM_Task} progress is 97%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.924340] env[65522]: WARNING openstack [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.924723] env[65522]: WARNING openstack [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.015041] env[65522]: DEBUG nova.scheduler.client.report [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 723.016726] env[65522]: DEBUG oslo_concurrency.lockutils [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.285s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 723.016887] env[65522]: INFO nova.compute.manager [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Successfully reverted task state from image_uploading on failure for instance. [ 723.021994] env[65522]: DEBUG nova.network.neutron [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Updated VIF entry in instance network info cache for port 37951fc0-2513-48ca-8e6a-42be7de65465. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 723.023926] env[65522]: DEBUG nova.network.neutron [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Updating instance_info_cache with network_info: [{"id": "37951fc0-2513-48ca-8e6a-42be7de65465", "address": "fa:16:3e:5e:53:01", "network": {"id": "922eeeeb-e124-4ae5-97a4-0e995b16f7ec", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-437359683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37fc01876567476f9b93d765b2cfddc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccf76700-491b-4462-ab19-e6d3a9ff87ac", "external-id": "nsx-vlan-transportzone-956", "segmentation_id": 956, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37951fc0-25", "ovs_interfaceid": "37951fc0-2513-48ca-8e6a-42be7de65465", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 723.024108] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.790s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 723.025555] env[65522]: INFO nova.compute.claims [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.034568] env[65522]: ERROR oslo_messaging.rpc.server [None req-06243bc2-c225-4644-b17c-45d250a13558 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Exception during message handling: glanceclient.exc.HTTPException: HTTP 410 Gone: Image 4cd7fa95-e6d9-457c-8cf1-e3881f933468 could not be found after upload. The image may have been deleted during the upload, cleaning up the chunks uploaded. [ 723.034568] env[65522]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 723.034568] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 723.034568] env[65522]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 723.034568] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 723.034568] env[65522]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 723.034568] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 723.034568] env[65522]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 723.034568] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 723.034568] env[65522]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 723.034568] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 723.034568] env[65522]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 723.034568] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 723.034568] env[65522]: ERROR oslo_messaging.rpc.server raise self.value [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 165, in decorated_function [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server raise self.value [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 156, in decorated_function [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 212, in decorated_function [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 723.034987] env[65522]: ERROR oslo_messaging.rpc.server raise self.value [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 202, in decorated_function [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 232, in decorated_function [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server raise self.value [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 229, in decorated_function [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server return function(self, context, image_id, instance, [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 4600, in snapshot_instance [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server self._snapshot_instance(context, image_id, instance, [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 4633, in _snapshot_instance [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server self.driver.snapshot(context, instance, image_id, [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 573, in snapshot [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server self._vmops.snapshot(context, instance, image_id, update_task_state) [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1024, in snapshot [ 723.035525] env[65522]: ERROR oslo_messaging.rpc.server images.upload_image_stream_optimized( [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/images.py", line 467, in upload_image_stream_optimized [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server IMAGE_API.update(context, image_id, image_metadata, data=read_handle) [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/image/glance.py", line 1254, in update [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server return session.update(context, image_id, image_info, data=data, [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/image/glance.py", line 704, in update [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server _reraise_translated_image_exception(image_id) [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/image/glance.py", line 1042, in _reraise_translated_image_exception [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server raise new_exc.with_traceback(exc_trace) [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/image/glance.py", line 702, in update [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server image = self._update_v2(context, sent_service_image_meta, data) [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/image/glance.py", line 721, in _update_v2 [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server image = self._upload_data(context, image_id, data) [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/image/glance.py", line 590, in _upload_data [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server utils.tpool_execute(self._client.call, [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/utils.py", line 635, in tpool_execute [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server return _pass_context(tpool.execute, func, *args, **kwargs) [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/utils.py", line 587, in _pass_context [ 723.035902] env[65522]: ERROR oslo_messaging.rpc.server return runner(context_wrapper, *args, **kwargs) [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/tpool.py", line 125, in execute [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server raise e.with_traceback(tb) [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/tpool.py", line 82, in tworker [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server rv = meth(*args, **kwargs) [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/utils.py", line 585, in context_wrapper [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server return func(*args, **kwargs) [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/image/glance.py", line 181, in call [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server result = getattr(controller, method)(*args, **kwargs) [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/glanceclient/common/utils.py", line 652, in inner [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server return RequestIdProxy(wrapped(*args, **kwargs)) [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/glanceclient/v2/images.py", line 313, in upload [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server resp, body = self.http_client.put(url, headers=hdrs, data=body) [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/keystoneauth1/adapter.py", line 619, in put [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server return self.request(url, 'PUT', **kwargs) [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/glanceclient/common/http.py", line 376, in request [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server return self._handle_response(resp) [ 723.036386] env[65522]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/glanceclient/common/http.py", line 116, in _handle_response [ 723.036824] env[65522]: ERROR oslo_messaging.rpc.server raise exc.from_response(resp, resp.content) [ 723.036824] env[65522]: ERROR oslo_messaging.rpc.server glanceclient.exc.HTTPException: HTTP 410 Gone: Image 4cd7fa95-e6d9-457c-8cf1-e3881f933468 could not be found after upload. The image may have been deleted during the upload, cleaning up the chunks uploaded. [ 723.036824] env[65522]: ERROR oslo_messaging.rpc.server [ 723.085848] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5113775, 'name': ReconfigVM_Task, 'duration_secs': 0.429703} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.086238] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Reconfigured VM instance instance-0000001e to attach disk [datastore1] 75c8848b-aea9-43f1-8697-9224050d1fef/75c8848b-aea9-43f1-8697-9224050d1fef.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 723.087828] env[65522]: WARNING neutronclient.v2_0.client [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.088537] env[65522]: WARNING openstack [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.088873] env[65522]: WARNING openstack [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.097346] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09a62cec-993a-474e-886d-93405ac002db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.112417] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 723.112417] env[65522]: value = "task-5113779" [ 723.112417] env[65522]: _type = "Task" [ 723.112417] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.114696] env[65522]: INFO nova.compute.manager [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Took 26.96 seconds to build instance. [ 723.141090] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5113779, 'name': Rename_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.143752] env[65522]: INFO nova.compute.manager [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Took 23.63 seconds to build instance. [ 723.153536] env[65522]: DEBUG oslo_vmware.api [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Task: {'id': task-5113777, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211893} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.154430] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 723.154430] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 723.154430] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 723.154430] env[65522]: INFO nova.compute.manager [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Took 1.25 seconds to destroy the instance on the hypervisor. [ 723.154788] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 723.155443] env[65522]: DEBUG nova.compute.manager [-] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 723.155553] env[65522]: DEBUG nova.network.neutron [-] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 723.155799] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.156421] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.156695] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.183203] env[65522]: DEBUG oslo_vmware.api [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113778, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.195349} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.183541] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 723.183742] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 723.184188] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 723.184188] env[65522]: INFO nova.compute.manager [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Took 1.20 seconds to destroy the instance on the hypervisor. [ 723.184442] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 723.184669] env[65522]: DEBUG nova.compute.manager [-] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 723.184783] env[65522]: DEBUG nova.network.neutron [-] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 723.185039] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.185577] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.185826] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.202196] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.265669] env[65522]: DEBUG nova.network.neutron [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Updating instance_info_cache with network_info: [{"id": "ede8edfa-842b-482d-8e7b-34bde13a02f8", "address": "fa:16:3e:f6:50:03", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapede8edfa-84", "ovs_interfaceid": "ede8edfa-842b-482d-8e7b-34bde13a02f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 723.344847] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.416842] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113766, 'name': RelocateVM_Task} progress is 98%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.426158] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquiring lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.426427] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquired lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 723.426867] env[65522]: WARNING neutronclient.v2_0.client [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.429751] env[65522]: WARNING openstack [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.429751] env[65522]: WARNING openstack [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.498474] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Acquiring lock "1d3a9ada-df46-46af-a7f1-a84c251af4ac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 723.498935] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lock "1d3a9ada-df46-46af-a7f1-a84c251af4ac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 723.499227] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Acquiring lock "1d3a9ada-df46-46af-a7f1-a84c251af4ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 723.499522] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lock "1d3a9ada-df46-46af-a7f1-a84c251af4ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 723.499810] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lock "1d3a9ada-df46-46af-a7f1-a84c251af4ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 723.502711] env[65522]: INFO nova.compute.manager [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Terminating instance [ 723.529884] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Releasing lock "refresh_cache-75c8848b-aea9-43f1-8697-9224050d1fef" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 723.530179] env[65522]: DEBUG nova.compute.manager [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Received event network-vif-plugged-d4e8efaa-adbc-4eec-adf0-e3f651352ba9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 723.530424] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Acquiring lock "42a08cfb-d865-4967-a086-370a2ca98b7d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 723.530572] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Lock "42a08cfb-d865-4967-a086-370a2ca98b7d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 723.530717] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Lock "42a08cfb-d865-4967-a086-370a2ca98b7d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 723.531080] env[65522]: DEBUG nova.compute.manager [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] No waiting events found dispatching network-vif-plugged-d4e8efaa-adbc-4eec-adf0-e3f651352ba9 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 723.531080] env[65522]: WARNING nova.compute.manager [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Received unexpected event network-vif-plugged-d4e8efaa-adbc-4eec-adf0-e3f651352ba9 for instance with vm_state building and task_state spawning. [ 723.531264] env[65522]: DEBUG nova.compute.manager [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Received event network-changed-d4e8efaa-adbc-4eec-adf0-e3f651352ba9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 723.531496] env[65522]: DEBUG nova.compute.manager [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Refreshing instance network info cache due to event network-changed-d4e8efaa-adbc-4eec-adf0-e3f651352ba9. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 723.531597] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Acquiring lock "refresh_cache-42a08cfb-d865-4967-a086-370a2ca98b7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.531733] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Acquired lock "refresh_cache-42a08cfb-d865-4967-a086-370a2ca98b7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 723.531899] env[65522]: DEBUG nova.network.neutron [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Refreshing network info cache for port d4e8efaa-adbc-4eec-adf0-e3f651352ba9 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 723.613963] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "57d32fb8-885e-44ba-967b-69f4bd1f5744" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 723.613963] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "57d32fb8-885e-44ba-967b-69f4bd1f5744" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 723.613963] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "57d32fb8-885e-44ba-967b-69f4bd1f5744-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 723.613963] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "57d32fb8-885e-44ba-967b-69f4bd1f5744-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 723.614259] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "57d32fb8-885e-44ba-967b-69f4bd1f5744-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 723.619880] env[65522]: INFO nova.compute.manager [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Terminating instance [ 723.622049] env[65522]: DEBUG oslo_concurrency.lockutils [None req-764df767-07a0-47d3-8b38-5c7b5d995407 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Lock "6e10e846-cdb7-458e-b511-18ab1742228e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.495s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 723.630099] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5113779, 'name': Rename_Task, 'duration_secs': 0.396024} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.631117] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 723.631408] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bc13eabb-d177-4a15-8579-f1d934f2f1a4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.640641] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 723.640641] env[65522]: value = "task-5113780" [ 723.640641] env[65522]: _type = "Task" [ 723.640641] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.647337] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ab4db85d-f0f4-4524-bca5-5a005eb98c9e tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lock "c0c99008-1907-4d3a-b4f6-ee6e916e2c67" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.168s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 723.647887] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 723.648202] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08562b3e-89a1-436b-a813-30ac87173a30 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.655965] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5113780, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.657500] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 723.657500] env[65522]: value = "task-5113781" [ 723.657500] env[65522]: _type = "Task" [ 723.657500] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.666841] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113781, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.769055] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "refresh_cache-ca964440-5375-4aff-8b45-96fbe829dd16" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 723.786997] env[65522]: DEBUG nova.compute.manager [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 723.815225] env[65522]: DEBUG nova.virt.hardware [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 723.815638] env[65522]: DEBUG nova.virt.hardware [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 723.815898] env[65522]: DEBUG nova.virt.hardware [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 723.815962] env[65522]: DEBUG nova.virt.hardware [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 723.816137] env[65522]: DEBUG nova.virt.hardware [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 723.816294] env[65522]: DEBUG nova.virt.hardware [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 723.816560] env[65522]: DEBUG nova.virt.hardware [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.816731] env[65522]: DEBUG nova.virt.hardware [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 723.816894] env[65522]: DEBUG nova.virt.hardware [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 723.817061] env[65522]: DEBUG nova.virt.hardware [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 723.817229] env[65522]: DEBUG nova.virt.hardware [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 723.818564] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c0a0318-2d8c-44e5-b355-fc35503a3604 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.830575] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2556ab09-5ad2-4aef-b329-f4286c5055b6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.920085] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113766, 'name': RelocateVM_Task, 'duration_secs': 4.499178} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.920447] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 723.920723] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994729', 'volume_id': '78aedba8-1717-4210-afe1-44eb7ae91932', 'name': 'volume-78aedba8-1717-4210-afe1-44eb7ae91932', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '42a08cfb-d865-4967-a086-370a2ca98b7d', 'attached_at': '', 'detached_at': '', 'volume_id': '78aedba8-1717-4210-afe1-44eb7ae91932', 'serial': '78aedba8-1717-4210-afe1-44eb7ae91932'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 723.921961] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd4b72e-f7f0-460d-a51c-0a6bf419cacb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.942115] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec54b0db-85ef-4679-8883-56f0f6854b0a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.966381] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Reconfiguring VM instance instance-0000001f to attach disk [datastore2] volume-78aedba8-1717-4210-afe1-44eb7ae91932/volume-78aedba8-1717-4210-afe1-44eb7ae91932.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 723.968373] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b5b6195-b68a-4c66-8df9-d1225745afa0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.993788] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Waiting for the task: (returnval){ [ 723.993788] env[65522]: value = "task-5113782" [ 723.993788] env[65522]: _type = "Task" [ 723.993788] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.010400] env[65522]: DEBUG nova.compute.manager [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 724.010707] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 724.011095] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113782, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.012339] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161595d4-5840-4ee5-84a0-00764104806b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.022086] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 724.022304] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b255bc7b-883f-4640-8dcc-b9f19e9585b6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.030915] env[65522]: DEBUG oslo_vmware.api [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 724.030915] env[65522]: value = "task-5113783" [ 724.030915] env[65522]: _type = "Task" [ 724.030915] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.036043] env[65522]: WARNING neutronclient.v2_0.client [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 724.036818] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.037221] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.053035] env[65522]: DEBUG oslo_vmware.api [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113783, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.078865] env[65522]: DEBUG nova.network.neutron [-] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 724.125088] env[65522]: DEBUG nova.compute.manager [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 724.125348] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 724.126406] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e9af1e-baf0-44df-88c8-6b91abc5e177 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.144150] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 724.148037] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fed06560-27b1-459d-8ab3-bae094c47d6d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.161206] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5113780, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.167156] env[65522]: DEBUG oslo_vmware.api [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 724.167156] env[65522]: value = "task-5113784" [ 724.167156] env[65522]: _type = "Task" [ 724.167156] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.181409] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113781, 'name': PowerOffVM_Task, 'duration_secs': 0.210253} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.182484] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 724.182994] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 724.184269] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49efb3a6-aa9c-4efe-ab7d-f7ea55415b0a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.192593] env[65522]: DEBUG oslo_vmware.api [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113784, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.202221] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 724.202614] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1c90b8f4-d9d3-4f90-8ec2-72e7c71369c7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.301032] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 724.301032] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 724.301262] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleting the datastore file [datastore1] f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 724.303374] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-76664cd1-9cfc-4ec3-9be4-b3c8f0380679 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.314371] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 724.314371] env[65522]: value = "task-5113786" [ 724.314371] env[65522]: _type = "Task" [ 724.314371] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.330053] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113786, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.489019] env[65522]: DEBUG nova.network.neutron [-] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 724.517619] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113782, 'name': ReconfigVM_Task, 'duration_secs': 0.438643} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.518036] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Reconfigured VM instance instance-0000001f to attach disk [datastore2] volume-78aedba8-1717-4210-afe1-44eb7ae91932/volume-78aedba8-1717-4210-afe1-44eb7ae91932.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 724.529623] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a86bfb6-b335-44f6-bc0d-c822d7730ff3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.560347] env[65522]: DEBUG oslo_vmware.api [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113783, 'name': PowerOffVM_Task, 'duration_secs': 0.244435} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.565776] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 724.566216] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 724.566540] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Waiting for the task: (returnval){ [ 724.566540] env[65522]: value = "task-5113787" [ 724.566540] env[65522]: _type = "Task" [ 724.566540] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.567531] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6cf52dc7-e62a-4d55-95e3-b01c9d2a93eb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.580433] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113787, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.581649] env[65522]: DEBUG nova.network.neutron [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Successfully updated port: b7e73ab8-d1b6-4e32-845a-fb308037ddf9 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 724.584163] env[65522]: INFO nova.compute.manager [-] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Took 1.43 seconds to deallocate network for instance. [ 724.654638] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 724.654935] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 724.655197] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Deleting the datastore file [datastore2] 1d3a9ada-df46-46af-a7f1-a84c251af4ac {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 724.656029] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fdb899af-9f7d-4eaa-8dea-0877ad7de425 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.664597] env[65522]: DEBUG oslo_vmware.api [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5113780, 'name': PowerOnVM_Task, 'duration_secs': 0.582797} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.667039] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 724.667396] env[65522]: INFO nova.compute.manager [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Took 10.86 seconds to spawn the instance on the hypervisor. [ 724.667689] env[65522]: DEBUG nova.compute.manager [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 724.668968] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978b8f08-223d-4b2b-8a4a-24dfe60a389b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.675865] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69ba0ec2-21c8-43c4-b9d6-0ca40449df37 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.681505] env[65522]: DEBUG oslo_vmware.api [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for the task: (returnval){ [ 724.681505] env[65522]: value = "task-5113789" [ 724.681505] env[65522]: _type = "Task" [ 724.681505] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.707594] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1b93b2-a36f-48ee-b664-e4045119bcc2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.714424] env[65522]: DEBUG oslo_vmware.api [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113784, 'name': PowerOffVM_Task, 'duration_secs': 0.288299} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.721281] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.721856] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.732495] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 724.732742] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 724.733128] env[65522]: DEBUG oslo_vmware.api [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113789, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.734963] env[65522]: DEBUG nova.network.neutron [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 724.740834] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d7addf02-475b-4919-95ec-02c0a419c17c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.773171] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a34fd5d-3179-4f9a-a591-c18d37f39f4b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.783073] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421e6b10-0522-4d72-98db-a12b795c9647 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.800799] env[65522]: DEBUG nova.compute.provider_tree [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.819324] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 724.819530] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 724.819831] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Deleting the datastore file [datastore1] 57d32fb8-885e-44ba-967b-69f4bd1f5744 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 724.820982] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-33c9b3da-0434-4a7c-b631-c0f15bd2ca2a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.831770] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113786, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.330859} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.833498] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 724.834211] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 724.834211] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 724.836804] env[65522]: DEBUG oslo_vmware.api [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for the task: (returnval){ [ 724.836804] env[65522]: value = "task-5113791" [ 724.836804] env[65522]: _type = "Task" [ 724.836804] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.847117] env[65522]: DEBUG oslo_vmware.api [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113791, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.852133] env[65522]: WARNING neutronclient.v2_0.client [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 724.852826] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.853181] env[65522]: WARNING openstack [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.968032] env[65522]: DEBUG nova.network.neutron [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Updated VIF entry in instance network info cache for port d4e8efaa-adbc-4eec-adf0-e3f651352ba9. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 724.968494] env[65522]: DEBUG nova.network.neutron [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Updating instance_info_cache with network_info: [{"id": "d4e8efaa-adbc-4eec-adf0-e3f651352ba9", "address": "fa:16:3e:b9:0a:a7", "network": {"id": "2bb5d3a6-95f2-43ba-9606-ad864f78a377", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-2069424406-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d64b94285334dd6b26928cabedbd67c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4e8efaa-ad", "ovs_interfaceid": "d4e8efaa-adbc-4eec-adf0-e3f651352ba9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 724.988943] env[65522]: INFO nova.compute.manager [-] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Took 1.80 seconds to deallocate network for instance. [ 725.081052] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113787, 'name': ReconfigVM_Task, 'duration_secs': 0.186715} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.081365] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994729', 'volume_id': '78aedba8-1717-4210-afe1-44eb7ae91932', 'name': 'volume-78aedba8-1717-4210-afe1-44eb7ae91932', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '42a08cfb-d865-4967-a086-370a2ca98b7d', 'attached_at': '', 'detached_at': '', 'volume_id': '78aedba8-1717-4210-afe1-44eb7ae91932', 'serial': '78aedba8-1717-4210-afe1-44eb7ae91932'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 725.081960] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0970e7b9-5336-4129-a075-b4b0d56ecc4b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.090319] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Waiting for the task: (returnval){ [ 725.090319] env[65522]: value = "task-5113792" [ 725.090319] env[65522]: _type = "Task" [ 725.090319] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.095175] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Acquiring lock "refresh_cache-20855931-a02d-4b53-b998-3d89d4ff5d1a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.095308] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Acquired lock "refresh_cache-20855931-a02d-4b53-b998-3d89d4ff5d1a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 725.095538] env[65522]: DEBUG nova.network.neutron [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 725.098902] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 725.105335] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113792, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.199179] env[65522]: DEBUG oslo_vmware.api [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Task: {'id': task-5113789, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1905} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.199443] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 725.199655] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 725.199905] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 725.200130] env[65522]: INFO nova.compute.manager [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Took 1.19 seconds to destroy the instance on the hypervisor. [ 725.200434] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 725.200668] env[65522]: DEBUG nova.compute.manager [-] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 725.200775] env[65522]: DEBUG nova.network.neutron [-] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 725.201150] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.201708] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.201967] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.224677] env[65522]: INFO nova.compute.manager [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Took 17.96 seconds to build instance. [ 725.243035] env[65522]: WARNING neutronclient.v2_0.client [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.244100] env[65522]: WARNING openstack [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.244100] env[65522]: WARNING openstack [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.273032] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.313713] env[65522]: DEBUG nova.scheduler.client.report [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 725.321453] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 725.321863] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-498469e7-1706-4d5e-9071-bbc4e222a7dd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.331068] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 725.331068] env[65522]: value = "task-5113793" [ 725.331068] env[65522]: _type = "Task" [ 725.331068] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.351398] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113793, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.362216] env[65522]: DEBUG oslo_vmware.api [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Task: {'id': task-5113791, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181513} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.362596] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 725.362900] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 725.363155] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 725.363386] env[65522]: INFO nova.compute.manager [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Took 1.24 seconds to destroy the instance on the hypervisor. [ 725.363694] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 725.363954] env[65522]: DEBUG nova.compute.manager [-] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 725.363954] env[65522]: DEBUG nova.network.neutron [-] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 725.364416] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.365035] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.365356] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.451352] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.471466] env[65522]: DEBUG oslo_concurrency.lockutils [req-c3bfc185-b611-4d90-9b6f-d7369509589b req-a6bdd7f8-398d-4e0d-9dfb-17e94a81ff31 service nova] Releasing lock "refresh_cache-42a08cfb-d865-4967-a086-370a2ca98b7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 725.496632] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 725.607192] env[65522]: WARNING openstack [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.607747] env[65522]: WARNING openstack [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.616361] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113792, 'name': Rename_Task, 'duration_secs': 0.398791} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.618108] env[65522]: WARNING openstack [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.618572] env[65522]: WARNING openstack [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.625576] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 725.626542] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1047e906-e767-47d4-98f4-9a753d4c9eac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.635376] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Waiting for the task: (returnval){ [ 725.635376] env[65522]: value = "task-5113794" [ 725.635376] env[65522]: _type = "Task" [ 725.635376] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.645877] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113794, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.692181] env[65522]: DEBUG nova.network.neutron [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 725.727500] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f72c2809-8d7d-481d-80b7-ba5997504dfe tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "75c8848b-aea9-43f1-8697-9224050d1fef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.485s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 725.747512] env[65522]: WARNING openstack [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.747908] env[65522]: WARNING openstack [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.808457] env[65522]: WARNING neutronclient.v2_0.client [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.808457] env[65522]: WARNING openstack [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.808837] env[65522]: WARNING openstack [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.820178] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.796s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 725.821071] env[65522]: DEBUG nova.compute.manager [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 725.825495] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.318s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 725.827145] env[65522]: INFO nova.compute.claims [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 725.830772] env[65522]: DEBUG nova.compute.manager [req-529d96f2-d48d-491d-902d-5d81f593a850 req-226de664-b4f8-49d1-9970-5505ba346bb8 service nova] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Received event network-vif-deleted-40d1ea0d-4c4d-4d05-9af8-ba1637b96c97 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 725.847051] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113793, 'name': PowerOffVM_Task, 'duration_secs': 0.304101} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.854627] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 725.860655] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04020491-d009-4dcd-a576-5af2aed086ca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.890718] env[65522]: WARNING neutronclient.v2_0.client [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.891429] env[65522]: WARNING openstack [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.891723] env[65522]: WARNING openstack [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.902077] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd642a8a-7c6c-4f8d-888a-2507fc53f01b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.920342] env[65522]: DEBUG nova.virt.hardware [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 725.920724] env[65522]: DEBUG nova.virt.hardware [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 725.920809] env[65522]: DEBUG nova.virt.hardware [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 725.920977] env[65522]: DEBUG nova.virt.hardware [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 725.921137] env[65522]: DEBUG nova.virt.hardware [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 725.921279] env[65522]: DEBUG nova.virt.hardware [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 725.921479] env[65522]: DEBUG nova.virt.hardware [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.921631] env[65522]: DEBUG nova.virt.hardware [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 725.921795] env[65522]: DEBUG nova.virt.hardware [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 725.921975] env[65522]: DEBUG nova.virt.hardware [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 725.922195] env[65522]: DEBUG nova.virt.hardware [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 725.923675] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8fe1cb9-b0a3-41e4-b0f8-ed3e6ded13d7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.940286] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9246037-24ba-4cfb-9ca0-823175b392df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.956232] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:53:d6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6fb0104-186b-4288-b87e-634893f46f01', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '189bde05-0b1d-4c6a-be88-2453345dd731', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 725.964405] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 725.966979] env[65522]: DEBUG nova.network.neutron [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Updating instance_info_cache with network_info: [{"id": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "address": "fa:16:3e:9b:22:d2", "network": {"id": "b1e15367-b861-4a6c-8f94-921ed2b0a66f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1304672897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a47babe780ec4ae7b6c40f4531f09a38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "79c2e589-f55b-4843-8d99-2e565be16706", "external-id": "nsx-vlan-transportzone-858", "segmentation_id": 858, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcf081b2-7a", "ovs_interfaceid": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 725.976024] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 725.976024] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 725.976024] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-628e2b5f-8d5e-4246-8825-651e5f295b06 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.977278] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32d92cd9-c399-4e44-9447-323a1b798f65 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.999929] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 725.999929] env[65522]: value = "task-5113795" [ 725.999929] env[65522]: _type = "Task" [ 725.999929] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.000923] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 726.000923] env[65522]: value = "task-5113796" [ 726.000923] env[65522]: _type = "Task" [ 726.000923] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.019670] env[65522]: DEBUG nova.network.neutron [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Updating instance_info_cache with network_info: [{"id": "b7e73ab8-d1b6-4e32-845a-fb308037ddf9", "address": "fa:16:3e:eb:5d:53", "network": {"id": "86cdd89e-db0e-4917-91d2-38f72500c93f", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1017903076-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "fd5d40deeddf430baae1517fa72e2214", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d", "external-id": "nsx-vlan-transportzone-591", "segmentation_id": 591, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7e73ab8-d1", "ovs_interfaceid": "b7e73ab8-d1b6-4e32-845a-fb308037ddf9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 726.021091] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 726.021300] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 726.021554] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.021750] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 726.021946] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 726.022191] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113796, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.022712] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-426a18d4-94ff-4216-a776-f51a7fd45696 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.025132] env[65522]: DEBUG nova.network.neutron [-] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 726.038840] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 726.039045] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 726.040135] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e054ea50-f16e-4ac6-89a1-a39b0cc035b5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.047331] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 726.047331] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527a1f9a-8249-ef1e-dccf-ed52556b437a" [ 726.047331] env[65522]: _type = "Task" [ 726.047331] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.056462] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527a1f9a-8249-ef1e-dccf-ed52556b437a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.147399] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113794, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.251327] env[65522]: DEBUG nova.compute.manager [req-c704350c-24ee-4792-b042-5b15764aa4a3 req-4dc6749e-60ca-4985-adab-6ebb579ddc72 service nova] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Received event network-vif-plugged-b7e73ab8-d1b6-4e32-845a-fb308037ddf9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 726.251574] env[65522]: DEBUG oslo_concurrency.lockutils [req-c704350c-24ee-4792-b042-5b15764aa4a3 req-4dc6749e-60ca-4985-adab-6ebb579ddc72 service nova] Acquiring lock "20855931-a02d-4b53-b998-3d89d4ff5d1a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 726.251824] env[65522]: DEBUG oslo_concurrency.lockutils [req-c704350c-24ee-4792-b042-5b15764aa4a3 req-4dc6749e-60ca-4985-adab-6ebb579ddc72 service nova] Lock "20855931-a02d-4b53-b998-3d89d4ff5d1a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 726.252299] env[65522]: DEBUG oslo_concurrency.lockutils [req-c704350c-24ee-4792-b042-5b15764aa4a3 req-4dc6749e-60ca-4985-adab-6ebb579ddc72 service nova] Lock "20855931-a02d-4b53-b998-3d89d4ff5d1a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.252594] env[65522]: DEBUG nova.compute.manager [req-c704350c-24ee-4792-b042-5b15764aa4a3 req-4dc6749e-60ca-4985-adab-6ebb579ddc72 service nova] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] No waiting events found dispatching network-vif-plugged-b7e73ab8-d1b6-4e32-845a-fb308037ddf9 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 726.252781] env[65522]: WARNING nova.compute.manager [req-c704350c-24ee-4792-b042-5b15764aa4a3 req-4dc6749e-60ca-4985-adab-6ebb579ddc72 service nova] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Received unexpected event network-vif-plugged-b7e73ab8-d1b6-4e32-845a-fb308037ddf9 for instance with vm_state building and task_state spawning. [ 726.270557] env[65522]: INFO nova.compute.manager [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Rebuilding instance [ 726.327750] env[65522]: DEBUG nova.compute.manager [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 726.328914] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eaae059-8d11-46ce-b25d-d17ee1ac47eb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.336602] env[65522]: DEBUG nova.compute.utils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 726.338731] env[65522]: DEBUG nova.compute.manager [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 726.338731] env[65522]: DEBUG nova.network.neutron [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 726.339127] env[65522]: WARNING neutronclient.v2_0.client [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 726.339643] env[65522]: WARNING neutronclient.v2_0.client [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 726.340607] env[65522]: WARNING openstack [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 726.341206] env[65522]: WARNING openstack [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 726.350853] env[65522]: DEBUG nova.compute.manager [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 726.365410] env[65522]: DEBUG nova.network.neutron [-] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 726.418056] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ae5297ea-0900-45f8-8f50-9af003734807 tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Acquiring lock "interface-c0c99008-1907-4d3a-b4f6-ee6e916e2c67-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 726.418327] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ae5297ea-0900-45f8-8f50-9af003734807 tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lock "interface-c0c99008-1907-4d3a-b4f6-ee6e916e2c67-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 726.418797] env[65522]: DEBUG nova.objects.instance [None req-ae5297ea-0900-45f8-8f50-9af003734807 tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lazy-loading 'flavor' on Instance uuid c0c99008-1907-4d3a-b4f6-ee6e916e2c67 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 726.432926] env[65522]: DEBUG nova.policy [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95f5aad7f373489fba64cf98b8b31ebc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfbeddbba72d447d93831a4b381f0ea0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 726.479062] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Releasing lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 726.479062] env[65522]: DEBUG nova.compute.manager [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Inject network info {{(pid=65522) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7794}} [ 726.479423] env[65522]: DEBUG nova.compute.manager [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] network_info to inject: |[{"id": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "address": "fa:16:3e:9b:22:d2", "network": {"id": "b1e15367-b861-4a6c-8f94-921ed2b0a66f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1304672897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a47babe780ec4ae7b6c40f4531f09a38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "79c2e589-f55b-4843-8d99-2e565be16706", "external-id": "nsx-vlan-transportzone-858", "segmentation_id": 858, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcf081b2-7a", "ovs_interfaceid": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7795}} [ 726.484643] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Reconfiguring VM instance to set the machine id {{(pid=65522) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 726.484994] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a219287-c96d-4ef5-92c4-9bfbd5b762a7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.503206] env[65522]: DEBUG oslo_vmware.api [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 726.503206] env[65522]: value = "task-5113797" [ 726.503206] env[65522]: _type = "Task" [ 726.503206] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.516045] env[65522]: DEBUG oslo_vmware.api [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113797, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.519732] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113796, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.524034] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Releasing lock "refresh_cache-20855931-a02d-4b53-b998-3d89d4ff5d1a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 726.524415] env[65522]: DEBUG nova.compute.manager [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Instance network_info: |[{"id": "b7e73ab8-d1b6-4e32-845a-fb308037ddf9", "address": "fa:16:3e:eb:5d:53", "network": {"id": "86cdd89e-db0e-4917-91d2-38f72500c93f", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1017903076-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "fd5d40deeddf430baae1517fa72e2214", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d", "external-id": "nsx-vlan-transportzone-591", "segmentation_id": 591, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7e73ab8-d1", "ovs_interfaceid": "b7e73ab8-d1b6-4e32-845a-fb308037ddf9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 726.524886] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:5d:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b7e73ab8-d1b6-4e32-845a-fb308037ddf9', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 726.533049] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Creating folder: Project (fd5d40deeddf430baae1517fa72e2214). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 726.533567] env[65522]: INFO nova.compute.manager [-] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Took 1.33 seconds to deallocate network for instance. [ 726.533825] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-adafa42e-7c6c-4dd6-9290-56a71fb63943 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.547408] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Created folder: Project (fd5d40deeddf430baae1517fa72e2214) in parent group-v994660. [ 726.547600] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Creating folder: Instances. Parent ref: group-v994751. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 726.547914] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ee25b698-7a06-465e-a311-738d6787e379 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.560299] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527a1f9a-8249-ef1e-dccf-ed52556b437a, 'name': SearchDatastore_Task, 'duration_secs': 0.020535} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.561778] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a706297-706e-4e7d-a130-9b85cfb8a839 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.565644] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Created folder: Instances in parent group-v994751. [ 726.565920] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 726.566153] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 726.566404] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e4642cd-8f3e-42b5-b761-53e94b108e9c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.584704] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 726.584704] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52667a35-f04f-6ee2-4f7a-488469afd788" [ 726.584704] env[65522]: _type = "Task" [ 726.584704] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.592034] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 726.592034] env[65522]: value = "task-5113800" [ 726.592034] env[65522]: _type = "Task" [ 726.592034] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.598507] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52667a35-f04f-6ee2-4f7a-488469afd788, 'name': SearchDatastore_Task, 'duration_secs': 0.01455} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.598844] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 726.599167] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] ca964440-5375-4aff-8b45-96fbe829dd16/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk. {{(pid=65522) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 726.599928] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ee2a6f0-7482-44f8-b30a-54f90d09009b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.606348] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113800, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.612786] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 726.612786] env[65522]: value = "task-5113801" [ 726.612786] env[65522]: _type = "Task" [ 726.612786] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.621767] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113801, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.648936] env[65522]: DEBUG oslo_vmware.api [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5113794, 'name': PowerOnVM_Task, 'duration_secs': 0.713483} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.649291] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 726.649554] env[65522]: INFO nova.compute.manager [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Took 8.48 seconds to spawn the instance on the hypervisor. [ 726.649779] env[65522]: DEBUG nova.compute.manager [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 726.650751] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffed2dc-502d-4013-8e48-7a7ae3f62477 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.870964] env[65522]: INFO nova.compute.manager [-] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Took 1.51 seconds to deallocate network for instance. [ 726.926295] env[65522]: WARNING neutronclient.v2_0.client [None req-ae5297ea-0900-45f8-8f50-9af003734807 tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 726.926774] env[65522]: WARNING openstack [None req-ae5297ea-0900-45f8-8f50-9af003734807 tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 726.927666] env[65522]: WARNING openstack [None req-ae5297ea-0900-45f8-8f50-9af003734807 tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 726.934373] env[65522]: DEBUG nova.objects.instance [None req-ae5297ea-0900-45f8-8f50-9af003734807 tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lazy-loading 'pci_requests' on Instance uuid c0c99008-1907-4d3a-b4f6-ee6e916e2c67 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 726.937026] env[65522]: DEBUG nova.network.neutron [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Successfully created port: 0345f514-e2bb-40e4-9692-c753d1dad222 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 727.036025] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113796, 'name': CreateVM_Task, 'duration_secs': 0.811875} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.036474] env[65522]: DEBUG oslo_vmware.api [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113797, 'name': ReconfigVM_Task, 'duration_secs': 0.243771} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.039863] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 727.039863] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ebdcdfb6-4ce1-47a3-93e8-b41b7b5084da tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Reconfigured VM instance to set the machine id {{(pid=65522) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 727.043726] env[65522]: WARNING neutronclient.v2_0.client [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 727.043726] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.043726] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 727.043726] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 727.043961] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.044494] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-823019ad-bda7-4851-8804-34eee29a726f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.053157] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 727.053157] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e753ca-183b-42ad-c66b-a156771af884" [ 727.053157] env[65522]: _type = "Task" [ 727.053157] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.069489] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e753ca-183b-42ad-c66b-a156771af884, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.106925] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113800, 'name': CreateVM_Task, 'duration_secs': 0.488027} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.110649] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 727.111717] env[65522]: WARNING neutronclient.v2_0.client [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 727.112232] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.125243] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113801, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.171958] env[65522]: INFO nova.compute.manager [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Took 17.45 seconds to build instance. [ 727.366093] env[65522]: DEBUG nova.compute.manager [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 727.368841] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 727.372135] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2271b156-7a08-418a-8a92-bdb6d07becfb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.381952] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.382514] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 727.382514] env[65522]: value = "task-5113802" [ 727.382514] env[65522]: _type = "Task" [ 727.382514] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.399167] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113802, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.409283] env[65522]: DEBUG nova.virt.hardware [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 727.410047] env[65522]: DEBUG nova.virt.hardware [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 727.410047] env[65522]: DEBUG nova.virt.hardware [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 727.410182] env[65522]: DEBUG nova.virt.hardware [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 727.410335] env[65522]: DEBUG nova.virt.hardware [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 727.410501] env[65522]: DEBUG nova.virt.hardware [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 727.410719] env[65522]: DEBUG nova.virt.hardware [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 727.410873] env[65522]: DEBUG nova.virt.hardware [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 727.411044] env[65522]: DEBUG nova.virt.hardware [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 727.411207] env[65522]: DEBUG nova.virt.hardware [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 727.411375] env[65522]: DEBUG nova.virt.hardware [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 727.412490] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412721a3-c834-435b-abd3-30babdd0f9ff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.421330] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb5a901-d1fd-4097-97e7-ce566717b31d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.442773] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b06fbd-0f85-4cb6-9835-852c002b211d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.446352] env[65522]: DEBUG nova.objects.base [None req-ae5297ea-0900-45f8-8f50-9af003734807 tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 727.446931] env[65522]: DEBUG nova.network.neutron [None req-ae5297ea-0900-45f8-8f50-9af003734807 tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 727.447087] env[65522]: WARNING neutronclient.v2_0.client [None req-ae5297ea-0900-45f8-8f50-9af003734807 tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 727.447606] env[65522]: WARNING neutronclient.v2_0.client [None req-ae5297ea-0900-45f8-8f50-9af003734807 tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 727.448140] env[65522]: WARNING openstack [None req-ae5297ea-0900-45f8-8f50-9af003734807 tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.448567] env[65522]: WARNING openstack [None req-ae5297ea-0900-45f8-8f50-9af003734807 tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.463026] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec9482f-79fe-49af-9786-a1b58ea1dbb7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.500383] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe69072f-8e0a-4ce0-944a-a3b1a37f581d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.509065] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c08fa749-faa5-4250-b79a-8e5653a68427 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.524905] env[65522]: DEBUG nova.compute.provider_tree [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.564810] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e753ca-183b-42ad-c66b-a156771af884, 'name': SearchDatastore_Task, 'duration_secs': 0.067404} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.565158] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 727.565422] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 727.565660] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.565807] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 727.565981] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 727.566690] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 727.566782] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 727.566973] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29f60912-c7bb-4d0a-a23e-7a7f669399d0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.569664] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f4bcfb7-a7fd-42c3-b964-d055c3f91323 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.576348] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Waiting for the task: (returnval){ [ 727.576348] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5235874c-0418-02fe-ba41-edc15d9abcf6" [ 727.576348] env[65522]: _type = "Task" [ 727.576348] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.580269] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 727.580445] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 727.581515] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe22e918-d303-4e32-8b3a-8811e9e15c79 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.586576] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5235874c-0418-02fe-ba41-edc15d9abcf6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.590243] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 727.590243] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b96d74-3ae9-3b35-9a18-1f1d561d3850" [ 727.590243] env[65522]: _type = "Task" [ 727.590243] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.598452] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b96d74-3ae9-3b35-9a18-1f1d561d3850, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.624302] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113801, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.612508} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.624628] env[65522]: INFO nova.virt.vmwareapi.ds_util [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] ca964440-5375-4aff-8b45-96fbe829dd16/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk. [ 727.625307] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3dc2a76-e029-4bc0-9c5c-c2aa0b8d9395 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.650473] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Reconfiguring VM instance instance-00000014 to attach disk [datastore2] ca964440-5375-4aff-8b45-96fbe829dd16/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 727.650787] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-947bdb2a-f738-4a32-afcf-98132ea968c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.670730] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 727.670730] env[65522]: value = "task-5113803" [ 727.670730] env[65522]: _type = "Task" [ 727.670730] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.674685] env[65522]: DEBUG oslo_concurrency.lockutils [None req-898cf7a8-257d-4bad-bebf-0982c2bc56cc tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Lock "42a08cfb-d865-4967-a086-370a2ca98b7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.965s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.680786] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113803, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.800189] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ae5297ea-0900-45f8-8f50-9af003734807 tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lock "interface-c0c99008-1907-4d3a-b4f6-ee6e916e2c67-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.382s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.894754] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113802, 'name': PowerOffVM_Task, 'duration_secs': 0.151494} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.895147] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 727.895342] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 727.896125] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c54d32-c6d4-4de4-a128-dc363e20304c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.904049] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 727.904378] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e563f717-c99a-426f-8404-a3d489391a9b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.931932] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 727.933046] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 727.933046] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Deleting the datastore file [datastore1] 6e10e846-cdb7-458e-b511-18ab1742228e {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 727.933046] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9d0ef92-f6f4-48b9-9107-785d311516a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.942136] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 727.942136] env[65522]: value = "task-5113805" [ 727.942136] env[65522]: _type = "Task" [ 727.942136] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.954774] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113805, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.028701] env[65522]: DEBUG nova.scheduler.client.report [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 728.090473] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5235874c-0418-02fe-ba41-edc15d9abcf6, 'name': SearchDatastore_Task, 'duration_secs': 0.027264} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.090945] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.091191] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 728.091409] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.102115] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b96d74-3ae9-3b35-9a18-1f1d561d3850, 'name': SearchDatastore_Task, 'duration_secs': 0.042983} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.102908] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d5657d3-4bf5-4251-b302-272ae008fde3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.109544] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 728.109544] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dc9df4-95d3-04b9-5597-3597355d67ad" [ 728.109544] env[65522]: _type = "Task" [ 728.109544] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.118971] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dc9df4-95d3-04b9-5597-3597355d67ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.181158] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113803, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.454946] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113805, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.400404} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.455275] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 728.456505] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 728.456505] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.535036] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.709s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 728.535310] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 728.537914] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.969s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.539394] env[65522]: INFO nova.compute.claims [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.622630] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dc9df4-95d3-04b9-5597-3597355d67ad, 'name': SearchDatastore_Task, 'duration_secs': 0.025538} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.622925] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.623212] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7/f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 728.623968] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 728.623968] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 728.623968] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-814992d0-dd03-4ca5-b50e-4bf12d4f74f4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.627011] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4f84c5a1-048c-46eb-b669-38ef99ce4934 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.634347] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 728.634347] env[65522]: value = "task-5113806" [ 728.634347] env[65522]: _type = "Task" [ 728.634347] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.640271] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 728.640498] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 728.642065] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b36d25c5-9d45-468b-a904-ec6c27559728 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.647793] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113806, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.653710] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Waiting for the task: (returnval){ [ 728.653710] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5202a142-5141-e8af-2ab7-5d2276c86521" [ 728.653710] env[65522]: _type = "Task" [ 728.653710] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.664871] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5202a142-5141-e8af-2ab7-5d2276c86521, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.682475] env[65522]: DEBUG nova.network.neutron [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Successfully updated port: 0345f514-e2bb-40e4-9692-c753d1dad222 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 728.688806] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113803, 'name': ReconfigVM_Task, 'duration_secs': 0.796336} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.689296] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Reconfigured VM instance instance-00000014 to attach disk [datastore2] ca964440-5375-4aff-8b45-96fbe829dd16/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 728.690690] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e794e39b-77de-46cf-a466-1386ff89853e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.728413] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eeac0de3-d5e2-48f8-a11e-78b12f90cc18 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.748397] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 728.748397] env[65522]: value = "task-5113807" [ 728.748397] env[65522]: _type = "Task" [ 728.748397] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.765287] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113807, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.044172] env[65522]: DEBUG nova.compute.utils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 729.050019] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 729.051362] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 729.051767] env[65522]: WARNING neutronclient.v2_0.client [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.052119] env[65522]: WARNING neutronclient.v2_0.client [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.053133] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.053390] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.114108] env[65522]: DEBUG nova.policy [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fea95beae5204764bed110847f92907f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '29c2f8cc328d4f6b985d258442c440b8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 729.146037] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113806, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.166938] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5202a142-5141-e8af-2ab7-5d2276c86521, 'name': SearchDatastore_Task, 'duration_secs': 0.016552} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.167892] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1898cbe4-3754-425a-a0be-eff2cad3edd2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.176026] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Waiting for the task: (returnval){ [ 729.176026] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520b88eb-facb-bb5a-5e5d-d4ad2be186dd" [ 729.176026] env[65522]: _type = "Task" [ 729.176026] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.187132] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "refresh_cache-361e54c7-4309-48b3-ad3e-b2ebc783741c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.187132] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquired lock "refresh_cache-361e54c7-4309-48b3-ad3e-b2ebc783741c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 729.187280] env[65522]: DEBUG nova.network.neutron [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 729.188941] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520b88eb-facb-bb5a-5e5d-d4ad2be186dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.263783] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113807, 'name': ReconfigVM_Task, 'duration_secs': 0.355457} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.264106] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 729.265129] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01554a6d-96d5-43dc-af33-67db80a44c46 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.274335] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 729.274335] env[65522]: value = "task-5113808" [ 729.274335] env[65522]: _type = "Task" [ 729.274335] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.287365] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113808, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.493262] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Successfully created port: 847d52c1-cfc5-4051-9969-5b8b567f330b {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 729.506163] env[65522]: DEBUG nova.virt.hardware [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 729.506408] env[65522]: DEBUG nova.virt.hardware [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 729.506603] env[65522]: DEBUG nova.virt.hardware [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 729.506806] env[65522]: DEBUG nova.virt.hardware [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 729.507016] env[65522]: DEBUG nova.virt.hardware [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 729.507124] env[65522]: DEBUG nova.virt.hardware [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 729.507297] env[65522]: DEBUG nova.virt.hardware [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 729.507450] env[65522]: DEBUG nova.virt.hardware [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 729.507615] env[65522]: DEBUG nova.virt.hardware [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 729.507770] env[65522]: DEBUG nova.virt.hardware [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 729.508233] env[65522]: DEBUG nova.virt.hardware [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 729.508978] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad4b280-2555-4b61-9a16-3962b02d5422 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.519774] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3b4807-507c-4f71-8e54-eda93df9716e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.535641] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 729.542864] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 729.543105] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 729.543340] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4db7236e-a7be-41ea-afd6-ef77042309dd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.561255] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 729.580739] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 729.580739] env[65522]: value = "task-5113809" [ 729.580739] env[65522]: _type = "Task" [ 729.580739] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.594649] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113809, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.651515] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113806, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.666916} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.652304] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7/f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 729.652722] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 729.653641] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6a8e9e60-65e8-47df-8d7d-b24624893113 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.663469] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 729.663469] env[65522]: value = "task-5113810" [ 729.663469] env[65522]: _type = "Task" [ 729.663469] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.687746] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113810, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.692249] env[65522]: WARNING openstack [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.692249] env[65522]: WARNING openstack [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.705375] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520b88eb-facb-bb5a-5e5d-d4ad2be186dd, 'name': SearchDatastore_Task, 'duration_secs': 0.075875} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.709176] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 729.709527] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 20855931-a02d-4b53-b998-3d89d4ff5d1a/20855931-a02d-4b53-b998-3d89d4ff5d1a.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 729.714020] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a30b490a-df87-4c3e-8849-1ff5cdf58b54 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.721424] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Waiting for the task: (returnval){ [ 729.721424] env[65522]: value = "task-5113811" [ 729.721424] env[65522]: _type = "Task" [ 729.721424] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.737322] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113811, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.786728] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113808, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.096445] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113809, 'name': CreateVM_Task, 'duration_secs': 0.396057} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.096761] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 730.097271] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.097403] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 730.097904] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 730.102587] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1a87d92-2838-425b-a6d0-637c20c6abaf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.110898] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 730.110898] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52609278-073a-abe9-775d-8b695ef89fd8" [ 730.110898] env[65522]: _type = "Task" [ 730.110898] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.126374] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52609278-073a-abe9-775d-8b695ef89fd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.158688] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5edc7ebe-5757-428b-a4a3-ac5681e6cea3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.175504] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7334d765-9bd3-4e7f-ac5e-c1c08ec1d9fc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.188246] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113810, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084625} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.217724] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 730.220372] env[65522]: DEBUG nova.compute.manager [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Received event network-vif-deleted-2eec2687-b0a0-442d-8306-27ab523bf504 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 730.220619] env[65522]: DEBUG nova.compute.manager [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Received event network-changed-dcf081b2-7a3e-4efd-84fa-bbf035dfd690 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 730.220822] env[65522]: DEBUG nova.compute.manager [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Refreshing instance network info cache due to event network-changed-dcf081b2-7a3e-4efd-84fa-bbf035dfd690. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 730.221085] env[65522]: DEBUG oslo_concurrency.lockutils [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] Acquiring lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.221246] env[65522]: DEBUG oslo_concurrency.lockutils [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] Acquired lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 730.221416] env[65522]: DEBUG nova.network.neutron [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Refreshing network info cache for port dcf081b2-7a3e-4efd-84fa-bbf035dfd690 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 730.229427] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7396fc34-15f8-4c92-9639-0d496555ee32 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.233986] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3286f99-687e-4877-bc46-19921aa7e3bf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.283260] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00af5667-f1b9-46bb-99a2-429a7406f054 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.304039] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Reconfiguring VM instance instance-00000018 to attach disk [datastore2] f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7/f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 730.304039] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113811, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.305254] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-480b4030-5c2d-40b7-a891-4ef145f73b95 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.336125] env[65522]: DEBUG nova.compute.provider_tree [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.342843] env[65522]: DEBUG oslo_vmware.api [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113808, 'name': PowerOnVM_Task, 'duration_secs': 0.669764} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.343208] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 730.343208] env[65522]: value = "task-5113812" [ 730.343208] env[65522]: _type = "Task" [ 730.343208] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.345511] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 730.351454] env[65522]: DEBUG nova.compute.manager [None req-5ebffc3e-00ed-415f-9740-a8178fd37d24 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 730.352427] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d38dc8f-cc22-412f-9686-52061f818705 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.358480] env[65522]: DEBUG nova.network.neutron [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 730.369452] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113812, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.584159] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 730.612974] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 730.613292] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 730.613482] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 730.613725] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 730.613860] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 730.614040] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 730.614288] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 730.614469] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 730.614665] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 730.614834] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 730.615073] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 730.616078] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09af6937-006b-4c29-8203-712c66ce212f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.632456] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9897f7b2-2e6e-47ed-88a7-72e815a163b9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.637100] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52609278-073a-abe9-775d-8b695ef89fd8, 'name': SearchDatastore_Task, 'duration_secs': 0.03908} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.637434] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 730.637671] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 730.637987] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.638161] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 730.638339] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 730.639077] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c769dad3-4f9d-4f11-abac-1e0d0a55e30b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.653520] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 730.653948] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 730.654514] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23167b43-61df-42ed-b12d-d6447af229f4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.660549] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 730.660549] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5253fdd3-40c7-ff7b-206c-3aea8a70eb25" [ 730.660549] env[65522]: _type = "Task" [ 730.660549] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.669373] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5253fdd3-40c7-ff7b-206c-3aea8a70eb25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.726990] env[65522]: WARNING neutronclient.v2_0.client [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 730.728464] env[65522]: WARNING openstack [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.729056] env[65522]: WARNING openstack [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.748766] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113811, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.714433} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.749049] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 20855931-a02d-4b53-b998-3d89d4ff5d1a/20855931-a02d-4b53-b998-3d89d4ff5d1a.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 730.749260] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 730.749519] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c691693b-aea0-4990-852f-c5f0bd55b44a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.757805] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Waiting for the task: (returnval){ [ 730.757805] env[65522]: value = "task-5113813" [ 730.757805] env[65522]: _type = "Task" [ 730.757805] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.769852] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113813, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.844745] env[65522]: DEBUG nova.scheduler.client.report [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 730.858125] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113812, 'name': ReconfigVM_Task, 'duration_secs': 0.387896} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.858125] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Reconfigured VM instance instance-00000018 to attach disk [datastore2] f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7/f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 730.859042] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0a32e6c9-2a0f-4bfa-a82f-5ac4fb910d83 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.866070] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 730.866070] env[65522]: value = "task-5113814" [ 730.866070] env[65522]: _type = "Task" [ 730.866070] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.877062] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113814, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.138466] env[65522]: WARNING openstack [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 731.139127] env[65522]: WARNING openstack [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 731.170982] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5253fdd3-40c7-ff7b-206c-3aea8a70eb25, 'name': SearchDatastore_Task, 'duration_secs': 0.035816} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.171923] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f21e4f59-ae55-4e10-9e7c-40eef0f6c480 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.180130] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 731.180130] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528c35d8-63eb-ef6c-2bb8-3a54d683ddc2" [ 731.180130] env[65522]: _type = "Task" [ 731.180130] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.188103] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Successfully updated port: 847d52c1-cfc5-4051-9969-5b8b567f330b {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 731.193921] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528c35d8-63eb-ef6c-2bb8-3a54d683ddc2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.268772] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113813, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.16278} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.269380] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 731.270717] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6664c82e-0499-4f51-bb88-d95463d07d0d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.298632] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] 20855931-a02d-4b53-b998-3d89d4ff5d1a/20855931-a02d-4b53-b998-3d89d4ff5d1a.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 731.299323] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-550def78-6fcd-4896-a4e6-483f29390938 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.333020] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Waiting for the task: (returnval){ [ 731.333020] env[65522]: value = "task-5113815" [ 731.333020] env[65522]: _type = "Task" [ 731.333020] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.344290] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113815, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.357098] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.815s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 731.357098] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 731.359101] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.733s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 731.360841] env[65522]: INFO nova.compute.claims [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.384731] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113814, 'name': Rename_Task, 'duration_secs': 0.161334} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.387360] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 731.387360] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7297ec1a-6988-4735-a6d5-e5adebe6c218 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.399157] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 731.399157] env[65522]: value = "task-5113816" [ 731.399157] env[65522]: _type = "Task" [ 731.399157] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.407696] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113816, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.647785] env[65522]: DEBUG nova.compute.manager [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Received event network-changed-b7e73ab8-d1b6-4e32-845a-fb308037ddf9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 731.647785] env[65522]: DEBUG nova.compute.manager [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Refreshing instance network info cache due to event network-changed-b7e73ab8-d1b6-4e32-845a-fb308037ddf9. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 731.647785] env[65522]: DEBUG oslo_concurrency.lockutils [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] Acquiring lock "refresh_cache-20855931-a02d-4b53-b998-3d89d4ff5d1a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.647953] env[65522]: DEBUG oslo_concurrency.lockutils [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] Acquired lock "refresh_cache-20855931-a02d-4b53-b998-3d89d4ff5d1a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 731.648122] env[65522]: DEBUG nova.network.neutron [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Refreshing network info cache for port b7e73ab8-d1b6-4e32-845a-fb308037ddf9 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 731.692135] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528c35d8-63eb-ef6c-2bb8-3a54d683ddc2, 'name': SearchDatastore_Task, 'duration_secs': 0.0566} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.692135] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 731.692690] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 6e10e846-cdb7-458e-b511-18ab1742228e/6e10e846-cdb7-458e-b511-18ab1742228e.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 731.692690] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c617a86d-7a68-4268-a906-0fd2b68901e3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.695528] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "refresh_cache-428873b8-1e58-4924-86f3-8feca8f53342" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.695751] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquired lock "refresh_cache-428873b8-1e58-4924-86f3-8feca8f53342" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 731.695967] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 731.705009] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 731.705009] env[65522]: value = "task-5113817" [ 731.705009] env[65522]: _type = "Task" [ 731.705009] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.718567] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113817, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.832602] env[65522]: WARNING neutronclient.v2_0.client [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 731.833225] env[65522]: WARNING openstack [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 731.833851] env[65522]: WARNING openstack [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 731.851468] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113815, 'name': ReconfigVM_Task, 'duration_secs': 0.364763} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.851763] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Reconfigured VM instance instance-00000020 to attach disk [datastore2] 20855931-a02d-4b53-b998-3d89d4ff5d1a/20855931-a02d-4b53-b998-3d89d4ff5d1a.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 731.852439] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-527793a0-8586-4311-94cf-ff22d28d1cf0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.862418] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Waiting for the task: (returnval){ [ 731.862418] env[65522]: value = "task-5113818" [ 731.862418] env[65522]: _type = "Task" [ 731.862418] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.868851] env[65522]: DEBUG nova.compute.utils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 731.877405] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 731.877405] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 731.877405] env[65522]: WARNING neutronclient.v2_0.client [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 731.877405] env[65522]: WARNING neutronclient.v2_0.client [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 731.877405] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 731.877776] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 731.897253] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113818, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.909914] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113816, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.152615] env[65522]: WARNING neutronclient.v2_0.client [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 732.153796] env[65522]: WARNING openstack [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.154459] env[65522]: WARNING openstack [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.199161] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.199878] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.221282] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113817, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.351882] env[65522]: DEBUG nova.policy [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fea95beae5204764bed110847f92907f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '29c2f8cc328d4f6b985d258442c440b8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 732.376369] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113818, 'name': Rename_Task, 'duration_secs': 0.175307} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.376799] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 732.377161] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5fcde131-47d9-4ae0-8a92-0e92969bc7f5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.391871] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 732.396799] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Waiting for the task: (returnval){ [ 732.396799] env[65522]: value = "task-5113819" [ 732.396799] env[65522]: _type = "Task" [ 732.396799] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.399034] env[65522]: WARNING openstack [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.399421] env[65522]: WARNING openstack [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.424519] env[65522]: DEBUG oslo_vmware.api [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113816, 'name': PowerOnVM_Task, 'duration_secs': 0.881701} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.428671] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 732.429116] env[65522]: DEBUG nova.compute.manager [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 732.429920] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113819, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.435770] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed209c6a-fd41-4117-a8f6-2a4265c465fb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.454341] env[65522]: DEBUG nova.network.neutron [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Updating instance_info_cache with network_info: [{"id": "0345f514-e2bb-40e4-9692-c753d1dad222", "address": "fa:16:3e:ef:a3:77", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.158", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0345f514-e2", "ovs_interfaceid": "0345f514-e2bb-40e4-9692-c753d1dad222", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 732.591296] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 732.727625] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113817, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.806042} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.732319] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 6e10e846-cdb7-458e-b511-18ab1742228e/6e10e846-cdb7-458e-b511-18ab1742228e.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 732.732319] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 732.733492] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e59a4b24-8dae-44a3-84e2-cfa702250ec7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.742305] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 732.742305] env[65522]: value = "task-5113820" [ 732.742305] env[65522]: _type = "Task" [ 732.742305] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.756631] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113820, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.925453] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113819, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.963030] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Releasing lock "refresh_cache-361e54c7-4309-48b3-ad3e-b2ebc783741c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 732.963528] env[65522]: DEBUG nova.compute.manager [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Instance network_info: |[{"id": "0345f514-e2bb-40e4-9692-c753d1dad222", "address": "fa:16:3e:ef:a3:77", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.158", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0345f514-e2", "ovs_interfaceid": "0345f514-e2bb-40e4-9692-c753d1dad222", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 732.966145] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:a3:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0345f514-e2bb-40e4-9692-c753d1dad222', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 732.975429] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Creating folder: Project (bfbeddbba72d447d93831a4b381f0ea0). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 732.976328] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 732.979498] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f175caba-1747-4730-a87f-e9c6daa7ecab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.991049] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Created folder: Project (bfbeddbba72d447d93831a4b381f0ea0) in parent group-v994660. [ 732.991250] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Creating folder: Instances. Parent ref: group-v994755. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 732.991494] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7122fa5-fc17-4dd4-bf8d-b59840284191 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.006485] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Created folder: Instances in parent group-v994755. [ 733.006884] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 733.009685] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 733.010155] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e907fc0-60e5-41b4-b807-ca10fcd41188 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.036047] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 733.036047] env[65522]: value = "task-5113823" [ 733.036047] env[65522]: _type = "Task" [ 733.036047] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.049855] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113823, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.051661] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-096a4f4c-b85f-4714-8539-831c4e5b8792 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.062148] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e252e6-1dc5-4ce3-bdd1-f55d3f09a1bd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.102985] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Successfully created port: 95b7c754-4f56-4bea-84f7-f4269d730b51 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 733.105995] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6363fde8-949a-42fc-b631-dd4b7ee1298e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.115215] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68a4b52-fc0f-4da2-be7d-ceaa8a0d4236 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.133725] env[65522]: DEBUG nova.compute.provider_tree [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.181117] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 733.181524] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 733.256307] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113820, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086098} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.256481] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 733.257673] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e8e7ad-569c-4724-a6a0-5ee82460f74f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.281906] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] 6e10e846-cdb7-458e-b511-18ab1742228e/6e10e846-cdb7-458e-b511-18ab1742228e.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 733.282301] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-060d6a16-857f-4f2e-8c8e-0a5e8385af41 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.306143] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 733.306143] env[65522]: value = "task-5113824" [ 733.306143] env[65522]: _type = "Task" [ 733.306143] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.318218] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113824, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.404856] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 733.417644] env[65522]: DEBUG oslo_vmware.api [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113819, 'name': PowerOnVM_Task, 'duration_secs': 0.56022} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.420052] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 733.420052] env[65522]: INFO nova.compute.manager [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Took 9.63 seconds to spawn the instance on the hypervisor. [ 733.420052] env[65522]: DEBUG nova.compute.manager [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 733.420052] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260c2977-38b9-49f7-b893-1c5169a80db2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.424118] env[65522]: WARNING neutronclient.v2_0.client [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 733.424924] env[65522]: WARNING openstack [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 733.425294] env[65522]: WARNING openstack [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 733.448830] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 733.449164] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 733.449317] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 733.449638] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 733.449638] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 733.449844] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 733.450102] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 733.450262] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 733.450422] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 733.450577] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 733.450740] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 733.453019] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f75d9a5d-6e46-4a0f-bafa-14720141efad {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.461183] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f5a46b-4fc1-40df-ae48-69317616fceb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.550036] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113823, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.637521] env[65522]: DEBUG nova.scheduler.client.report [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 733.650857] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Acquiring lock "c0c99008-1907-4d3a-b4f6-ee6e916e2c67" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.651183] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lock "c0c99008-1907-4d3a-b4f6-ee6e916e2c67" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.651582] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Acquiring lock "c0c99008-1907-4d3a-b4f6-ee6e916e2c67-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.651932] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lock "c0c99008-1907-4d3a-b4f6-ee6e916e2c67-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.652211] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lock "c0c99008-1907-4d3a-b4f6-ee6e916e2c67-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 733.655826] env[65522]: INFO nova.compute.manager [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Terminating instance [ 733.817687] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113824, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.948912] env[65522]: INFO nova.compute.manager [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Took 14.99 seconds to build instance. [ 734.049332] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113823, 'name': CreateVM_Task, 'duration_secs': 0.739942} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.049874] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 734.050226] env[65522]: WARNING neutronclient.v2_0.client [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 734.050663] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.050935] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 734.051302] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 734.051590] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc475fb9-ce34-4ce2-a335-032d7481abb9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.057380] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 734.057380] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527fa497-c967-aa2d-2766-68a7de888b18" [ 734.057380] env[65522]: _type = "Task" [ 734.057380] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.074206] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527fa497-c967-aa2d-2766-68a7de888b18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.077136] env[65522]: WARNING openstack [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 734.078889] env[65522]: WARNING openstack [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 734.145615] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.786s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.146508] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 734.152090] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 13.792s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.162452] env[65522]: DEBUG nova.compute.manager [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 734.162664] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 734.163913] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d24bc85-29b7-4b35-b6fd-34dd0364cc4f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.177135] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 734.177422] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb343ed7-fadc-4f46-8c35-489dea672365 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.189270] env[65522]: DEBUG oslo_vmware.api [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Waiting for the task: (returnval){ [ 734.189270] env[65522]: value = "task-5113825" [ 734.189270] env[65522]: _type = "Task" [ 734.189270] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.203047] env[65522]: DEBUG oslo_vmware.api [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113825, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.324981] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113824, 'name': ReconfigVM_Task, 'duration_secs': 0.986614} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.325403] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Reconfigured VM instance instance-0000001c to attach disk [datastore1] 6e10e846-cdb7-458e-b511-18ab1742228e/6e10e846-cdb7-458e-b511-18ab1742228e.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 734.326498] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3f4c65e2-ac5f-4e47-b51d-7187f5f61ee9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.334678] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 734.334678] env[65522]: value = "task-5113826" [ 734.334678] env[65522]: _type = "Task" [ 734.334678] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.357536] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113826, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.451642] env[65522]: DEBUG oslo_concurrency.lockutils [None req-346dec77-b5c1-4879-a52e-533fea785b65 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Lock "20855931-a02d-4b53-b998-3d89d4ff5d1a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.510s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.586126] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527fa497-c967-aa2d-2766-68a7de888b18, 'name': SearchDatastore_Task, 'duration_secs': 0.03775} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.586776] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 734.587270] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 734.587783] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.588194] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 734.590028] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 734.590028] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b71b4887-affb-43bf-a620-179651300cd5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.601097] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 734.601760] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 734.603098] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00ae585f-6e07-4115-99ff-6c58a9a48dee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.612393] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 734.612393] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5265d3da-0d2b-e930-f2f7-174731ad8175" [ 734.612393] env[65522]: _type = "Task" [ 734.612393] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.627389] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5265d3da-0d2b-e930-f2f7-174731ad8175, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.655066] env[65522]: DEBUG nova.compute.utils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 734.663310] env[65522]: INFO nova.compute.claims [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 734.668291] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 734.668291] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 734.668838] env[65522]: WARNING neutronclient.v2_0.client [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 734.669314] env[65522]: WARNING neutronclient.v2_0.client [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 734.670181] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 734.670943] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 734.678693] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 734.705183] env[65522]: DEBUG oslo_vmware.api [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113825, 'name': PowerOffVM_Task, 'duration_secs': 0.321108} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.705183] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 734.705183] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 734.705183] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3dbaa90d-b59f-4080-bd8b-157f6b99ae2c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.792087] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 734.792356] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 734.792563] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Deleting the datastore file [datastore1] c0c99008-1907-4d3a-b4f6-ee6e916e2c67 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 734.792876] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a3a4f90-f9dc-43df-a107-76a7ee074c31 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.803829] env[65522]: DEBUG oslo_vmware.api [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Waiting for the task: (returnval){ [ 734.803829] env[65522]: value = "task-5113828" [ 734.803829] env[65522]: _type = "Task" [ 734.803829] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.815158] env[65522]: DEBUG oslo_vmware.api [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113828, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.846080] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113826, 'name': Rename_Task, 'duration_secs': 0.257032} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.846398] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 734.846819] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e94c4102-ffc0-4f59-a2ad-35c51452b989 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.855238] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 734.855238] env[65522]: value = "task-5113829" [ 734.855238] env[65522]: _type = "Task" [ 734.855238] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.865614] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113829, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.114693] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Successfully updated port: 95b7c754-4f56-4bea-84f7-f4269d730b51 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 735.133517] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5265d3da-0d2b-e930-f2f7-174731ad8175, 'name': SearchDatastore_Task, 'duration_secs': 0.025412} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.135179] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fefa9cf-ab72-426e-b5c7-96aecbf213e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.145232] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 735.145232] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527eebb6-bf6a-021c-fe6b-3337b18a2c9a" [ 735.145232] env[65522]: _type = "Task" [ 735.145232] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.162019] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527eebb6-bf6a-021c-fe6b-3337b18a2c9a, 'name': SearchDatastore_Task, 'duration_secs': 0.013543} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.162373] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 735.162748] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 361e54c7-4309-48b3-ad3e-b2ebc783741c/361e54c7-4309-48b3-ad3e-b2ebc783741c.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 735.163146] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6bfd3ba6-10e0-4b78-bcc0-8890e82a6a72 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.172825] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 735.172825] env[65522]: value = "task-5113830" [ 735.172825] env[65522]: _type = "Task" [ 735.172825] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.183115] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113830, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.185497] env[65522]: INFO nova.compute.resource_tracker [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating resource usage from migration 86317211-4746-4c8e-8603-9fd37855968e [ 735.316073] env[65522]: DEBUG oslo_vmware.api [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Task: {'id': task-5113828, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.25186} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.316539] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 735.316789] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 735.317045] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.317414] env[65522]: INFO nova.compute.manager [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Took 1.15 seconds to destroy the instance on the hypervisor. [ 735.317679] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 735.318260] env[65522]: DEBUG nova.compute.manager [-] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 735.318260] env[65522]: DEBUG nova.network.neutron [-] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 735.318379] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 735.318954] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.319224] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.327895] env[65522]: WARNING neutronclient.v2_0.client [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 735.328549] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.328927] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.369580] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113829, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.506434] env[65522]: DEBUG nova.policy [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fea95beae5204764bed110847f92907f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '29c2f8cc328d4f6b985d258442c440b8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 735.621734] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "refresh_cache-54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.621734] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquired lock "refresh_cache-54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 735.621734] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 735.687231] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113830, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.700228] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 735.746281] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 735.747258] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 735.747587] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 735.748874] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 735.748874] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 735.748874] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 735.748874] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.749475] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 735.749721] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 735.750028] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 735.750497] env[65522]: DEBUG nova.virt.hardware [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 735.752127] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14abd7c8-f724-49ae-95bd-5cd1e3d884c9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.773471] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83df936f-1842-48b1-a81b-8325ed1c5e8f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.867420] env[65522]: DEBUG oslo_vmware.api [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113829, 'name': PowerOnVM_Task, 'duration_secs': 0.57122} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.868922] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 735.869337] env[65522]: DEBUG nova.compute.manager [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 735.871607] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9fc3db5-fd0f-4f4d-ba18-3f9b3db37213 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.875347] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f74ccef-9667-4ef4-b61b-1d10a3f095f4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.888846] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bdec7f6-ac83-408a-922e-3a6e788b6f26 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.926361] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae08a3d2-289c-438a-8eb3-888bb2ea0a80 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.935998] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c7eab5-e7e5-4482-b2f0-49abec69c54a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.954309] env[65522]: DEBUG nova.compute.provider_tree [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.127048] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 736.127503] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 736.159740] env[65522]: DEBUG nova.network.neutron [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Updated VIF entry in instance network info cache for port dcf081b2-7a3e-4efd-84fa-bbf035dfd690. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 736.159797] env[65522]: DEBUG nova.network.neutron [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Updating instance_info_cache with network_info: [{"id": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "address": "fa:16:3e:9b:22:d2", "network": {"id": "b1e15367-b861-4a6c-8f94-921ed2b0a66f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1304672897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a47babe780ec4ae7b6c40f4531f09a38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "79c2e589-f55b-4843-8d99-2e565be16706", "external-id": "nsx-vlan-transportzone-858", "segmentation_id": 858, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcf081b2-7a", "ovs_interfaceid": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 736.188154] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113830, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.393373] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 736.405531] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 736.456303] env[65522]: WARNING neutronclient.v2_0.client [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 736.457246] env[65522]: WARNING openstack [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 736.458219] env[65522]: WARNING openstack [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 736.467852] env[65522]: DEBUG nova.scheduler.client.report [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 736.633753] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 736.662394] env[65522]: DEBUG oslo_concurrency.lockutils [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] Releasing lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 736.662812] env[65522]: DEBUG nova.compute.manager [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Received event network-vif-deleted-3896379b-7eb4-4dc5-819d-f68421a44dab {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 736.663025] env[65522]: DEBUG nova.compute.manager [req-e747b30c-16ff-4516-9055-ffb6a254ce85 req-bc71178e-bfa5-4ae5-952b-e6cd20cbfea7 service nova] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Received event network-vif-deleted-7d49f270-6cbd-4767-91fc-f899c46564c9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 736.690669] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113830, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.874491] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Successfully created port: 93f5a0f6-2499-4475-b6e9-68d105b143f3 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 736.974414] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.825s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 736.975037] env[65522]: INFO nova.compute.manager [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Migrating [ 736.976711] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.976711] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired lock "compute-rpcapi-router" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 736.978156] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.081s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 736.980294] env[65522]: INFO nova.compute.claims [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.988429] env[65522]: INFO nova.compute.rpcapi [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Automatically selected compute RPC version 6.4 from minimum service version 70 [ 736.989671] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Releasing lock "compute-rpcapi-router" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 737.186818] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113830, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.193372] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Updating instance_info_cache with network_info: [{"id": "847d52c1-cfc5-4051-9969-5b8b567f330b", "address": "fa:16:3e:c6:f3:5e", "network": {"id": "f83e6c69-de0c-4306-ac88-4ad211e44d55", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1504496554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "29c2f8cc328d4f6b985d258442c440b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "392517be-6cb8-4b5b-9a52-449bfe2e16f7", "external-id": "nsx-vlan-transportzone-351", "segmentation_id": 351, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap847d52c1-cf", "ovs_interfaceid": "847d52c1-cfc5-4051-9969-5b8b567f330b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 737.286927] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 737.287659] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 737.506445] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.506599] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 737.506704] env[65522]: DEBUG nova.network.neutron [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 737.646496] env[65522]: DEBUG nova.network.neutron [-] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 737.660471] env[65522]: DEBUG nova.network.neutron [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Updated VIF entry in instance network info cache for port b7e73ab8-d1b6-4e32-845a-fb308037ddf9. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 737.660842] env[65522]: DEBUG nova.network.neutron [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Updating instance_info_cache with network_info: [{"id": "b7e73ab8-d1b6-4e32-845a-fb308037ddf9", "address": "fa:16:3e:eb:5d:53", "network": {"id": "86cdd89e-db0e-4917-91d2-38f72500c93f", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1017903076-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "fd5d40deeddf430baae1517fa72e2214", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7d09e9-a3dd-4d89-b9dd-2814f5f6dd5d", "external-id": "nsx-vlan-transportzone-591", "segmentation_id": 591, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7e73ab8-d1", "ovs_interfaceid": "b7e73ab8-d1b6-4e32-845a-fb308037ddf9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 737.690869] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113830, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.696195] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Releasing lock "refresh_cache-428873b8-1e58-4924-86f3-8feca8f53342" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 737.696618] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Instance network_info: |[{"id": "847d52c1-cfc5-4051-9969-5b8b567f330b", "address": "fa:16:3e:c6:f3:5e", "network": {"id": "f83e6c69-de0c-4306-ac88-4ad211e44d55", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1504496554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "29c2f8cc328d4f6b985d258442c440b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "392517be-6cb8-4b5b-9a52-449bfe2e16f7", "external-id": "nsx-vlan-transportzone-351", "segmentation_id": 351, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap847d52c1-cf", "ovs_interfaceid": "847d52c1-cfc5-4051-9969-5b8b567f330b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 737.697213] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:f3:5e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '392517be-6cb8-4b5b-9a52-449bfe2e16f7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '847d52c1-cfc5-4051-9969-5b8b567f330b', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 737.707939] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Creating folder: Project (29c2f8cc328d4f6b985d258442c440b8). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 737.708988] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f6e61427-1670-4feb-b628-b93380cd6c0d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.727690] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Created folder: Project (29c2f8cc328d4f6b985d258442c440b8) in parent group-v994660. [ 737.727968] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Creating folder: Instances. Parent ref: group-v994758. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 737.728328] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3b578bef-c727-430c-a608-2dd5c8e7d2e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.743713] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Created folder: Instances in parent group-v994758. [ 737.744080] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 737.745591] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 737.745591] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a083a37e-6686-4395-aef6-c7406dd608f1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.786573] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 737.786573] env[65522]: value = "task-5113833" [ 737.786573] env[65522]: _type = "Task" [ 737.786573] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.800589] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113833, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.009200] env[65522]: WARNING neutronclient.v2_0.client [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 738.011152] env[65522]: WARNING openstack [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 738.011152] env[65522]: WARNING openstack [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 738.152182] env[65522]: INFO nova.compute.manager [-] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Took 2.83 seconds to deallocate network for instance. [ 738.165263] env[65522]: DEBUG oslo_concurrency.lockutils [req-e01103d9-e95d-4b50-b9dd-fbfe71eb6245 req-8490c3e1-c642-4cd8-8d9b-1fc661b5e267 service nova] Releasing lock "refresh_cache-20855931-a02d-4b53-b998-3d89d4ff5d1a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 738.192236] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113830, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.933771} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.192236] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 361e54c7-4309-48b3-ad3e-b2ebc783741c/361e54c7-4309-48b3-ad3e-b2ebc783741c.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 738.192485] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 738.192655] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71edf45a-3600-408d-8cff-4055f24203f0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.204236] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 738.204236] env[65522]: value = "task-5113834" [ 738.204236] env[65522]: _type = "Task" [ 738.204236] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.215629] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113834, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.300339] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113833, 'name': CreateVM_Task, 'duration_secs': 0.487076} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.303633] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 738.303633] env[65522]: WARNING neutronclient.v2_0.client [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 738.303633] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.303633] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 738.303633] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 738.303633] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7dcd824e-fe23-441f-8272-55912ba9b251 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.311647] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 738.311647] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527adef9-98bd-a8fe-9bea-50a3814304f3" [ 738.311647] env[65522]: _type = "Task" [ 738.311647] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.323343] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527adef9-98bd-a8fe-9bea-50a3814304f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.498583] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Acquiring lock "ea5b8e29-ebe6-444b-bd2d-029201003ee6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 738.498831] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Lock "ea5b8e29-ebe6-444b-bd2d-029201003ee6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 738.649364] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1d8eb8-3054-4e83-9640-b28116a377df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.659740] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8723845f-6886-48a9-85c1-b7231b196cd6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.664420] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 738.695231] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5b768f-2af6-4780-81f5-da9d3ec1067a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.704602] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af77c385-7f26-481f-96f1-24f99a72b591 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.732838] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113834, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.170162} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.733488] env[65522]: DEBUG nova.compute.provider_tree [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.735222] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 738.736739] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd24050a-e052-4d65-af7f-d6ffd8b57850 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.761435] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] 361e54c7-4309-48b3-ad3e-b2ebc783741c/361e54c7-4309-48b3-ad3e-b2ebc783741c.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 738.762772] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ce61a48-a782-46ba-93bd-752dbf89445f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.786703] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 738.786703] env[65522]: value = "task-5113835" [ 738.786703] env[65522]: _type = "Task" [ 738.786703] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.798154] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113835, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.824263] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527adef9-98bd-a8fe-9bea-50a3814304f3, 'name': SearchDatastore_Task, 'duration_secs': 0.019682} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.824263] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 738.824263] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 738.824575] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.824644] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 738.825092] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 738.825791] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e704944a-f15e-45ed-a53b-c1c96d7737eb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.838458] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 738.838702] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 738.839806] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a474ab6-ebbe-4168-98c0-d63e76ca7317 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.847360] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 738.847360] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529857bb-8d75-e4c3-b8f9-fc24b99792f0" [ 738.847360] env[65522]: _type = "Task" [ 738.847360] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.859278] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529857bb-8d75-e4c3-b8f9-fc24b99792f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.957446] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Successfully updated port: 93f5a0f6-2499-4475-b6e9-68d105b143f3 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 739.006208] env[65522]: DEBUG nova.compute.manager [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 739.198764] env[65522]: WARNING neutronclient.v2_0.client [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 739.198764] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.198764] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.237692] env[65522]: DEBUG nova.scheduler.client.report [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 739.301758] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113835, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.360973] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529857bb-8d75-e4c3-b8f9-fc24b99792f0, 'name': SearchDatastore_Task, 'duration_secs': 0.027702} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.362170] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0fa4d94-c7d9-4d3c-a125-608f537d9843 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.369160] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 739.369160] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528ce273-e79c-4a14-18f3-bb0bc8abe42f" [ 739.369160] env[65522]: _type = "Task" [ 739.369160] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.382034] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528ce273-e79c-4a14-18f3-bb0bc8abe42f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.461254] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "refresh_cache-028c7fc5-5624-4dd5-9e2f-48191a86f765" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.461316] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquired lock "refresh_cache-028c7fc5-5624-4dd5-9e2f-48191a86f765" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 739.461481] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 739.550028] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 739.586802] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Updating instance_info_cache with network_info: [{"id": "95b7c754-4f56-4bea-84f7-f4269d730b51", "address": "fa:16:3e:30:82:2b", "network": {"id": "f83e6c69-de0c-4306-ac88-4ad211e44d55", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1504496554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "29c2f8cc328d4f6b985d258442c440b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "392517be-6cb8-4b5b-9a52-449bfe2e16f7", "external-id": "nsx-vlan-transportzone-351", "segmentation_id": 351, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b7c754-4f", "ovs_interfaceid": "95b7c754-4f56-4bea-84f7-f4269d730b51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 739.607732] env[65522]: WARNING openstack [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.608146] env[65522]: WARNING openstack [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.747402] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.770s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 739.748279] env[65522]: DEBUG nova.compute.manager [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 739.753056] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.655s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 739.753056] env[65522]: DEBUG nova.objects.instance [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lazy-loading 'resources' on Instance uuid dbf29bd1-780c-4756-a203-4fad05a9d3b9 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 739.807643] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113835, 'name': ReconfigVM_Task, 'duration_secs': 1.014476} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.809812] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Reconfigured VM instance instance-00000021 to attach disk [datastore2] 361e54c7-4309-48b3-ad3e-b2ebc783741c/361e54c7-4309-48b3-ad3e-b2ebc783741c.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 739.812015] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa9fe964-da67-4c8e-9216-1da63fdc8503 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.821839] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 739.821839] env[65522]: value = "task-5113836" [ 739.821839] env[65522]: _type = "Task" [ 739.821839] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.833117] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113836, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.848022] env[65522]: WARNING neutronclient.v2_0.client [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 739.849406] env[65522]: WARNING openstack [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.849946] env[65522]: WARNING openstack [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.887380] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528ce273-e79c-4a14-18f3-bb0bc8abe42f, 'name': SearchDatastore_Task, 'duration_secs': 0.029814} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.887665] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 739.888038] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 428873b8-1e58-4924-86f3-8feca8f53342/428873b8-1e58-4924-86f3-8feca8f53342.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 739.888357] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-69368249-5848-4e9e-b922-34f87eaddf9e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.896887] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 739.896887] env[65522]: value = "task-5113837" [ 739.896887] env[65522]: _type = "Task" [ 739.896887] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.906674] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113837, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.965553] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.966574] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.975203] env[65522]: DEBUG nova.objects.instance [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Lazy-loading 'flavor' on Instance uuid 6abf58ca-ad6b-4fe8-844a-ca33e62a844d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 740.071024] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 740.090775] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Releasing lock "refresh_cache-54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 740.091427] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Instance network_info: |[{"id": "95b7c754-4f56-4bea-84f7-f4269d730b51", "address": "fa:16:3e:30:82:2b", "network": {"id": "f83e6c69-de0c-4306-ac88-4ad211e44d55", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1504496554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "29c2f8cc328d4f6b985d258442c440b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "392517be-6cb8-4b5b-9a52-449bfe2e16f7", "external-id": "nsx-vlan-transportzone-351", "segmentation_id": 351, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b7c754-4f", "ovs_interfaceid": "95b7c754-4f56-4bea-84f7-f4269d730b51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 740.092583] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:82:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '392517be-6cb8-4b5b-9a52-449bfe2e16f7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '95b7c754-4f56-4bea-84f7-f4269d730b51', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 740.102611] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 740.104936] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 740.104936] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0f214995-3e70-4091-9e3c-5a11791fbb85 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.122552] env[65522]: DEBUG nova.network.neutron [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating instance_info_cache with network_info: [{"id": "fb1cd05a-757b-469c-9749-fd1ca4fcb168", "address": "fa:16:3e:b1:9d:ad", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb1cd05a-75", "ovs_interfaceid": "fb1cd05a-757b-469c-9749-fd1ca4fcb168", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 740.131330] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 740.131330] env[65522]: value = "task-5113838" [ 740.131330] env[65522]: _type = "Task" [ 740.131330] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.146764] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113838, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.198858] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.199506] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.257409] env[65522]: DEBUG nova.compute.utils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 740.266570] env[65522]: DEBUG nova.compute.manager [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 740.266570] env[65522]: DEBUG nova.network.neutron [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 740.266570] env[65522]: WARNING neutronclient.v2_0.client [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.266570] env[65522]: WARNING neutronclient.v2_0.client [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.266570] env[65522]: WARNING openstack [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.267221] env[65522]: WARNING openstack [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.341151] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113836, 'name': Rename_Task, 'duration_secs': 0.274682} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.341550] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 740.341925] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee44852e-bba4-4416-abc9-2fcccce30bda {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.355469] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 740.355469] env[65522]: value = "task-5113839" [ 740.355469] env[65522]: _type = "Task" [ 740.355469] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.372833] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113839, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.413400] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113837, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.488197] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquiring lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.488422] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquired lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 740.488865] env[65522]: WARNING neutronclient.v2_0.client [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.574386] env[65522]: DEBUG nova.compute.manager [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Received event network-changed-d4e8efaa-adbc-4eec-adf0-e3f651352ba9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 740.574386] env[65522]: DEBUG nova.compute.manager [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Refreshing instance network info cache due to event network-changed-d4e8efaa-adbc-4eec-adf0-e3f651352ba9. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 740.574731] env[65522]: DEBUG oslo_concurrency.lockutils [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] Acquiring lock "refresh_cache-42a08cfb-d865-4967-a086-370a2ca98b7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.575048] env[65522]: DEBUG oslo_concurrency.lockutils [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] Acquired lock "refresh_cache-42a08cfb-d865-4967-a086-370a2ca98b7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 740.575161] env[65522]: DEBUG nova.network.neutron [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Refreshing network info cache for port d4e8efaa-adbc-4eec-adf0-e3f651352ba9 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 740.626315] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Releasing lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 740.652539] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113838, 'name': CreateVM_Task, 'duration_secs': 0.498845} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.655528] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 740.656745] env[65522]: WARNING neutronclient.v2_0.client [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.658858] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.658858] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 740.658858] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 740.658858] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e4bf17d-209c-48c6-bfe6-f119e7f6b834 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.664648] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 740.664648] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c1d444-5f6a-0f87-8f97-3475c6f38697" [ 740.664648] env[65522]: _type = "Task" [ 740.664648] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.679999] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c1d444-5f6a-0f87-8f97-3475c6f38697, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.765152] env[65522]: WARNING neutronclient.v2_0.client [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.766046] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.766494] env[65522]: WARNING openstack [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.776440] env[65522]: DEBUG nova.compute.manager [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 740.825065] env[65522]: DEBUG nova.policy [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95f5aad7f373489fba64cf98b8b31ebc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfbeddbba72d447d93831a4b381f0ea0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 740.873858] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113839, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.910542] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113837, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.994434] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af48254-8e4c-427b-b42d-2070ea82fbbd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.010187] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb5184c-b4f1-41d3-ac5c-58442b553c52 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.053268] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d40108d-a932-42fc-82e3-5a9e38c4500f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.062125] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da802c07-cf2b-49da-adb5-62ff07e68928 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.080468] env[65522]: DEBUG nova.compute.provider_tree [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.083509] env[65522]: WARNING neutronclient.v2_0.client [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 741.084324] env[65522]: WARNING openstack [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 741.084735] env[65522]: WARNING openstack [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 741.177879] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c1d444-5f6a-0f87-8f97-3475c6f38697, 'name': SearchDatastore_Task, 'duration_secs': 0.030154} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.178187] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 741.178516] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 741.179910] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.179996] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 741.180226] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 741.180689] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2e8789cf-18a9-444a-97c4-ce2b2b2662df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.194708] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 741.195041] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 741.195871] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2596e95a-6bc0-44b5-9b50-d22812277407 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.203729] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 741.203729] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52172694-5325-6316-ec79-1413600f3fd8" [ 741.203729] env[65522]: _type = "Task" [ 741.203729] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.217519] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52172694-5325-6316-ec79-1413600f3fd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.245078] env[65522]: DEBUG nova.network.neutron [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 741.376185] env[65522]: DEBUG oslo_vmware.api [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113839, 'name': PowerOnVM_Task, 'duration_secs': 0.641727} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.376185] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 741.376185] env[65522]: INFO nova.compute.manager [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Took 14.01 seconds to spawn the instance on the hypervisor. [ 741.376185] env[65522]: DEBUG nova.compute.manager [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 741.376185] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eba6c21-a74b-48bd-a970-a52959421bc1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.403840] env[65522]: DEBUG nova.compute.manager [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Received event network-vif-plugged-0345f514-e2bb-40e4-9692-c753d1dad222 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 741.404971] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Acquiring lock "361e54c7-4309-48b3-ad3e-b2ebc783741c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 741.404971] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Lock "361e54c7-4309-48b3-ad3e-b2ebc783741c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 741.404971] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Lock "361e54c7-4309-48b3-ad3e-b2ebc783741c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 741.404971] env[65522]: DEBUG nova.compute.manager [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] No waiting events found dispatching network-vif-plugged-0345f514-e2bb-40e4-9692-c753d1dad222 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 741.404971] env[65522]: WARNING nova.compute.manager [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Received unexpected event network-vif-plugged-0345f514-e2bb-40e4-9692-c753d1dad222 for instance with vm_state building and task_state spawning. [ 741.405181] env[65522]: DEBUG nova.compute.manager [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Received event network-changed-0345f514-e2bb-40e4-9692-c753d1dad222 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 741.405373] env[65522]: DEBUG nova.compute.manager [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Refreshing instance network info cache due to event network-changed-0345f514-e2bb-40e4-9692-c753d1dad222. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 741.405594] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Acquiring lock "refresh_cache-361e54c7-4309-48b3-ad3e-b2ebc783741c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.405862] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Acquired lock "refresh_cache-361e54c7-4309-48b3-ad3e-b2ebc783741c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 741.406073] env[65522]: DEBUG nova.network.neutron [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Refreshing network info cache for port 0345f514-e2bb-40e4-9692-c753d1dad222 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 741.426042] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113837, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.460708] env[65522]: DEBUG nova.network.neutron [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Updating instance_info_cache with network_info: [{"id": "93f5a0f6-2499-4475-b6e9-68d105b143f3", "address": "fa:16:3e:f6:b8:ff", "network": {"id": "f83e6c69-de0c-4306-ac88-4ad211e44d55", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1504496554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "29c2f8cc328d4f6b985d258442c440b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "392517be-6cb8-4b5b-9a52-449bfe2e16f7", "external-id": "nsx-vlan-transportzone-351", "segmentation_id": 351, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93f5a0f6-24", "ovs_interfaceid": "93f5a0f6-2499-4475-b6e9-68d105b143f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 741.584136] env[65522]: DEBUG nova.scheduler.client.report [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 741.628928] env[65522]: WARNING openstack [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 741.630314] env[65522]: WARNING openstack [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 741.727028] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52172694-5325-6316-ec79-1413600f3fd8, 'name': SearchDatastore_Task, 'duration_secs': 0.081145} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.727637] env[65522]: DEBUG nova.network.neutron [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Successfully created port: 345f682f-4aaa-4bc1-bb61-e9d25e6288c5 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 741.734563] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19333c33-7ad6-4503-90af-6dffe836b38e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.745218] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 741.745218] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52128897-3f75-63ed-4e94-4e71175b8be1" [ 741.745218] env[65522]: _type = "Task" [ 741.745218] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.756427] env[65522]: WARNING neutronclient.v2_0.client [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 741.756737] env[65522]: WARNING openstack [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 741.757081] env[65522]: WARNING openstack [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 741.764613] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52128897-3f75-63ed-4e94-4e71175b8be1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.788972] env[65522]: DEBUG nova.compute.manager [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 741.826486] env[65522]: DEBUG nova.virt.hardware [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 741.826638] env[65522]: DEBUG nova.virt.hardware [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 741.826696] env[65522]: DEBUG nova.virt.hardware [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 741.826862] env[65522]: DEBUG nova.virt.hardware [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 741.826997] env[65522]: DEBUG nova.virt.hardware [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 741.827222] env[65522]: DEBUG nova.virt.hardware [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 741.827360] env[65522]: DEBUG nova.virt.hardware [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.827695] env[65522]: DEBUG nova.virt.hardware [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 741.827695] env[65522]: DEBUG nova.virt.hardware [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 741.827798] env[65522]: DEBUG nova.virt.hardware [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 741.827964] env[65522]: DEBUG nova.virt.hardware [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 741.829038] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da66215-ee65-44b9-bfb8-b562561c8ae4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.847025] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4090f8da-9128-4b10-8973-6accce2ef012 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.914090] env[65522]: INFO nova.compute.manager [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Took 22.70 seconds to build instance. [ 741.918451] env[65522]: WARNING neutronclient.v2_0.client [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 741.919539] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 741.919742] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 741.928124] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113837, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.77703} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.928569] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 428873b8-1e58-4924-86f3-8feca8f53342/428873b8-1e58-4924-86f3-8feca8f53342.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 741.928941] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 741.933496] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a13926e4-ff07-49af-ace0-edb6a18b6c1c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.937776] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 741.937776] env[65522]: value = "task-5113840" [ 741.937776] env[65522]: _type = "Task" [ 741.937776] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.950175] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113840, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.966732] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Releasing lock "refresh_cache-028c7fc5-5624-4dd5-9e2f-48191a86f765" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 741.967249] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Instance network_info: |[{"id": "93f5a0f6-2499-4475-b6e9-68d105b143f3", "address": "fa:16:3e:f6:b8:ff", "network": {"id": "f83e6c69-de0c-4306-ac88-4ad211e44d55", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1504496554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "29c2f8cc328d4f6b985d258442c440b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "392517be-6cb8-4b5b-9a52-449bfe2e16f7", "external-id": "nsx-vlan-transportzone-351", "segmentation_id": 351, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93f5a0f6-24", "ovs_interfaceid": "93f5a0f6-2499-4475-b6e9-68d105b143f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 741.967755] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:b8:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '392517be-6cb8-4b5b-9a52-449bfe2e16f7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '93f5a0f6-2499-4475-b6e9-68d105b143f3', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 741.976959] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 741.978147] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 741.978147] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ced791c7-38f7-4a3c-a75a-a9d3d44958cb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.000238] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 742.000238] env[65522]: value = "task-5113841" [ 742.000238] env[65522]: _type = "Task" [ 742.000238] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.009801] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113841, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.090230] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.337s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.093935] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.597s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.094538] env[65522]: DEBUG nova.objects.instance [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lazy-loading 'resources' on Instance uuid 6c8165f5-6769-4954-b9ac-de8551954377 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 742.125223] env[65522]: INFO nova.scheduler.client.report [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Deleted allocations for instance dbf29bd1-780c-4756-a203-4fad05a9d3b9 [ 742.136395] env[65522]: WARNING neutronclient.v2_0.client [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.136816] env[65522]: WARNING openstack [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.137161] env[65522]: WARNING openstack [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.154351] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee30b521-1701-4779-af98-66a1c253ca41 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.172907] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating instance 'd5e8dd05-dc3c-4831-b4b0-ac100360f3e7' progress to 0 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 742.257141] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52128897-3f75-63ed-4e94-4e71175b8be1, 'name': SearchDatastore_Task, 'duration_secs': 0.022131} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.257600] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 742.257658] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824/54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 742.257935] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eae29a9b-7446-488f-a136-4a0899fb14aa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.268035] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 742.268035] env[65522]: value = "task-5113842" [ 742.268035] env[65522]: _type = "Task" [ 742.268035] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.281027] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113842, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.428630] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4ff6afe3-06ce-41fe-a41c-fed397f0a6f3 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "361e54c7-4309-48b3-ad3e-b2ebc783741c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.233s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.450848] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113840, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082591} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.450848] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 742.453050] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c90e41e5-671d-41c5-9eee-24f7f90bc726 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.479480] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Reconfiguring VM instance instance-00000022 to attach disk [datastore1] 428873b8-1e58-4924-86f3-8feca8f53342/428873b8-1e58-4924-86f3-8feca8f53342.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 742.485342] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45d0a3a7-abdd-43fe-b0c9-f3b7a1e55634 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.515291] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113841, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.517028] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 742.517028] env[65522]: value = "task-5113843" [ 742.517028] env[65522]: _type = "Task" [ 742.517028] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.528423] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113843, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.607659] env[65522]: WARNING openstack [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.608073] env[65522]: WARNING openstack [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.639857] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dfa87759-720d-4175-9e49-5fe8ceda8640 tempest-DeleteServersAdminTestJSON-1353403214 tempest-DeleteServersAdminTestJSON-1353403214-project-member] Lock "dbf29bd1-780c-4756-a203-4fad05a9d3b9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.256s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.682838] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 742.683432] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3cad6d35-e138-4b51-9749-aa4045d201f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.697492] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 742.697492] env[65522]: value = "task-5113844" [ 742.697492] env[65522]: _type = "Task" [ 742.697492] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.709513] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113844, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.740352] env[65522]: DEBUG nova.network.neutron [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Updated VIF entry in instance network info cache for port d4e8efaa-adbc-4eec-adf0-e3f651352ba9. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 742.740352] env[65522]: DEBUG nova.network.neutron [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Updating instance_info_cache with network_info: [{"id": "d4e8efaa-adbc-4eec-adf0-e3f651352ba9", "address": "fa:16:3e:b9:0a:a7", "network": {"id": "2bb5d3a6-95f2-43ba-9606-ad864f78a377", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-2069424406-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6d64b94285334dd6b26928cabedbd67c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4e8efaa-ad", "ovs_interfaceid": "d4e8efaa-adbc-4eec-adf0-e3f651352ba9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 742.757050] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.757763] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.785527] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113842, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.922621] env[65522]: WARNING neutronclient.v2_0.client [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.923373] env[65522]: WARNING openstack [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.923852] env[65522]: WARNING openstack [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 743.026366] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113841, 'name': CreateVM_Task, 'duration_secs': 0.544143} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.027963] env[65522]: WARNING neutronclient.v2_0.client [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 743.028876] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 743.029586] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 743.037928] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 743.041824] env[65522]: WARNING neutronclient.v2_0.client [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 743.042051] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.042257] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 743.042611] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 743.047012] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1b7bc80-8420-4b2e-a98b-1650b1be34eb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.049758] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113843, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.053978] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 743.053978] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52af1bea-5d42-799c-aac1-62e7e97e652c" [ 743.053978] env[65522]: _type = "Task" [ 743.053978] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.066847] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52af1bea-5d42-799c-aac1-62e7e97e652c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.157610] env[65522]: DEBUG nova.network.neutron [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Updating instance_info_cache with network_info: [{"id": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "address": "fa:16:3e:9b:22:d2", "network": {"id": "b1e15367-b861-4a6c-8f94-921ed2b0a66f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1304672897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a47babe780ec4ae7b6c40f4531f09a38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "79c2e589-f55b-4843-8d99-2e565be16706", "external-id": "nsx-vlan-transportzone-858", "segmentation_id": 858, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcf081b2-7a", "ovs_interfaceid": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 743.212041] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113844, 'name': PowerOffVM_Task, 'duration_secs': 0.347022} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.212419] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 743.212525] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating instance 'd5e8dd05-dc3c-4831-b4b0-ac100360f3e7' progress to 17 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 743.225499] env[65522]: DEBUG nova.network.neutron [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Updated VIF entry in instance network info cache for port 0345f514-e2bb-40e4-9692-c753d1dad222. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 743.226565] env[65522]: DEBUG nova.network.neutron [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Updating instance_info_cache with network_info: [{"id": "0345f514-e2bb-40e4-9692-c753d1dad222", "address": "fa:16:3e:ef:a3:77", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.158", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0345f514-e2", "ovs_interfaceid": "0345f514-e2bb-40e4-9692-c753d1dad222", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 743.244146] env[65522]: DEBUG oslo_concurrency.lockutils [req-5324b9cc-7b87-435d-aacf-200a90a5aafc req-4a31b521-fc2d-4171-808a-2b8b21be4c46 service nova] Releasing lock "refresh_cache-42a08cfb-d865-4967-a086-370a2ca98b7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 743.279877] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d685441a-f860-4226-a6ad-8b60184f8d5b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.289478] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113842, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.745866} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.289478] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824/54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 743.289478] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 743.289478] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf9ca147-cd43-42a0-8b66-e3f59f42b3d7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.296180] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d0e743-3b87-4a95-a910-b03bef85ac5e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.301868] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 743.301868] env[65522]: value = "task-5113845" [ 743.301868] env[65522]: _type = "Task" [ 743.301868] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.338993] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c3719f-55a7-48b5-8508-1c9421183303 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.346283] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113845, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.352846] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90425825-8513-4998-922f-a7fef7891283 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.371473] env[65522]: DEBUG nova.compute.provider_tree [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.494256] env[65522]: DEBUG nova.network.neutron [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Successfully updated port: 345f682f-4aaa-4bc1-bb61-e9d25e6288c5 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 743.529649] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113843, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.569289] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52af1bea-5d42-799c-aac1-62e7e97e652c, 'name': SearchDatastore_Task, 'duration_secs': 0.015351} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.569781] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 743.570166] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 743.570537] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.570777] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 743.571092] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 743.571523] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8ff00fcc-d2fb-41ee-9d2c-13521485bca6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.593389] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 743.593676] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 743.594958] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-970393ae-40bd-489f-bd2a-9aee2a6f3618 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.602110] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 743.602110] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52423c9d-26ce-30b9-11bd-80c57a8502c0" [ 743.602110] env[65522]: _type = "Task" [ 743.602110] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.615980] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52423c9d-26ce-30b9-11bd-80c57a8502c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.661543] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Releasing lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 743.661625] env[65522]: DEBUG nova.compute.manager [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Inject network info {{(pid=65522) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7794}} [ 743.662011] env[65522]: DEBUG nova.compute.manager [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] network_info to inject: |[{"id": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "address": "fa:16:3e:9b:22:d2", "network": {"id": "b1e15367-b861-4a6c-8f94-921ed2b0a66f", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1304672897-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a47babe780ec4ae7b6c40f4531f09a38", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "79c2e589-f55b-4843-8d99-2e565be16706", "external-id": "nsx-vlan-transportzone-858", "segmentation_id": 858, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdcf081b2-7a", "ovs_interfaceid": "dcf081b2-7a3e-4efd-84fa-bbf035dfd690", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7795}} [ 743.666936] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Reconfiguring VM instance to set the machine id {{(pid=65522) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 743.667346] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a93494f8-5dd1-47c3-a871-60fd4bf40afb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.690184] env[65522]: DEBUG oslo_vmware.api [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 743.690184] env[65522]: value = "task-5113846" [ 743.690184] env[65522]: _type = "Task" [ 743.690184] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.704533] env[65522]: DEBUG oslo_vmware.api [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113846, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.723553] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 743.725023] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 743.725023] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 743.725023] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 743.725023] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 743.725023] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 743.725519] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 743.725899] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 743.726224] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 743.726566] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 743.727301] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 743.736122] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Releasing lock "refresh_cache-361e54c7-4309-48b3-ad3e-b2ebc783741c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 743.736122] env[65522]: DEBUG nova.compute.manager [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Received event network-vif-plugged-847d52c1-cfc5-4051-9969-5b8b567f330b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 743.736122] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Acquiring lock "428873b8-1e58-4924-86f3-8feca8f53342-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.736122] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Lock "428873b8-1e58-4924-86f3-8feca8f53342-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.736122] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Lock "428873b8-1e58-4924-86f3-8feca8f53342-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.736122] env[65522]: DEBUG nova.compute.manager [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] No waiting events found dispatching network-vif-plugged-847d52c1-cfc5-4051-9969-5b8b567f330b {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 743.736122] env[65522]: WARNING nova.compute.manager [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Received unexpected event network-vif-plugged-847d52c1-cfc5-4051-9969-5b8b567f330b for instance with vm_state building and task_state spawning. [ 743.736661] env[65522]: DEBUG nova.compute.manager [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Received event network-changed-847d52c1-cfc5-4051-9969-5b8b567f330b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 743.737064] env[65522]: DEBUG nova.compute.manager [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Refreshing instance network info cache due to event network-changed-847d52c1-cfc5-4051-9969-5b8b567f330b. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 743.737392] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Acquiring lock "refresh_cache-428873b8-1e58-4924-86f3-8feca8f53342" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.738105] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Acquired lock "refresh_cache-428873b8-1e58-4924-86f3-8feca8f53342" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 743.738105] env[65522]: DEBUG nova.network.neutron [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Refreshing network info cache for port 847d52c1-cfc5-4051-9969-5b8b567f330b {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 743.739453] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23599781-5f11-4c02-908f-2bf6464285f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.765800] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 743.765800] env[65522]: value = "task-5113847" [ 743.765800] env[65522]: _type = "Task" [ 743.765800] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.778349] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113847, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.818625] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113845, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08128} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.818625] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 743.818625] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f73d15e6-79cd-4394-8c6a-a35bcae9a6cb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.846146] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Reconfiguring VM instance instance-00000023 to attach disk [datastore1] 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824/54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 743.846579] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6225ce30-3e62-43b3-b0f2-6a4c4eefe80f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.869629] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "4238d766-2a5d-40d7-b2f6-5cb9a5354760" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.870021] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "4238d766-2a5d-40d7-b2f6-5cb9a5354760" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.877614] env[65522]: DEBUG nova.scheduler.client.report [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 743.881752] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 743.881752] env[65522]: value = "task-5113848" [ 743.881752] env[65522]: _type = "Task" [ 743.881752] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.895357] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113848, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.920292] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "1e8088ef-6354-40a0-95ee-06732e7f088f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.920292] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "1e8088ef-6354-40a0-95ee-06732e7f088f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.997290] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "refresh_cache-da211708-a6a2-4e03-bb41-c5b93564de20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.997430] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquired lock "refresh_cache-da211708-a6a2-4e03-bb41-c5b93564de20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 743.997611] env[65522]: DEBUG nova.network.neutron [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 744.031558] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113843, 'name': ReconfigVM_Task, 'duration_secs': 1.328675} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.033418] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Reconfigured VM instance instance-00000022 to attach disk [datastore1] 428873b8-1e58-4924-86f3-8feca8f53342/428873b8-1e58-4924-86f3-8feca8f53342.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 744.033418] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dabce34c-9232-4906-9cea-4b6e86667a2b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.043794] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 744.043794] env[65522]: value = "task-5113849" [ 744.043794] env[65522]: _type = "Task" [ 744.043794] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.058800] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113849, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.117663] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52423c9d-26ce-30b9-11bd-80c57a8502c0, 'name': SearchDatastore_Task, 'duration_secs': 0.030455} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.118546] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-112e3bc5-570c-4b71-a0a1-390f1f5f3dbf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.126619] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 744.126619] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52625f01-1f17-44e2-7e16-17d75d4b3ff4" [ 744.126619] env[65522]: _type = "Task" [ 744.126619] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.140884] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52625f01-1f17-44e2-7e16-17d75d4b3ff4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.204115] env[65522]: DEBUG oslo_vmware.api [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113846, 'name': ReconfigVM_Task, 'duration_secs': 0.206125} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.204489] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4998f34e-2639-422a-a8d4-e3b23d316f46 tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Reconfigured VM instance to set the machine id {{(pid=65522) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 744.261023] env[65522]: WARNING neutronclient.v2_0.client [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 744.261023] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.261023] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.282981] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113847, 'name': ReconfigVM_Task, 'duration_secs': 0.314291} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.286262] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating instance 'd5e8dd05-dc3c-4831-b4b0-ac100360f3e7' progress to 33 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 744.376703] env[65522]: DEBUG nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 744.383536] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.290s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.386799] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.342s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 744.386799] env[65522]: DEBUG nova.objects.instance [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lazy-loading 'resources' on Instance uuid 1d3a9ada-df46-46af-a7f1-a84c251af4ac {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 744.398774] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113848, 'name': ReconfigVM_Task, 'duration_secs': 0.310573} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.400201] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Reconfigured VM instance instance-00000023 to attach disk [datastore1] 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824/54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 744.402375] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3f3191ed-06a2-40c4-b32a-79ed302b4f29 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.415066] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 744.415066] env[65522]: value = "task-5113850" [ 744.415066] env[65522]: _type = "Task" [ 744.415066] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.426545] env[65522]: DEBUG nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 744.426834] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113850, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.427672] env[65522]: INFO nova.scheduler.client.report [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Deleted allocations for instance 6c8165f5-6769-4954-b9ac-de8551954377 [ 744.501125] env[65522]: WARNING openstack [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.501531] env[65522]: WARNING openstack [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.556767] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113849, 'name': Rename_Task, 'duration_secs': 0.18401} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.557078] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 744.557340] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65d838a8-be29-4e08-906d-d77a5ff94889 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.565517] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 744.565517] env[65522]: value = "task-5113851" [ 744.565517] env[65522]: _type = "Task" [ 744.565517] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.577615] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113851, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.639265] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52625f01-1f17-44e2-7e16-17d75d4b3ff4, 'name': SearchDatastore_Task, 'duration_secs': 0.011969} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.639553] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 744.639937] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 028c7fc5-5624-4dd5-9e2f-48191a86f765/028c7fc5-5624-4dd5-9e2f-48191a86f765.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 744.640119] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ee89348c-2690-400d-ae9b-4c49444ce7f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.648287] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 744.648287] env[65522]: value = "task-5113852" [ 744.648287] env[65522]: _type = "Task" [ 744.648287] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.654727] env[65522]: DEBUG nova.network.neutron [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 744.662586] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113852, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.792639] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 744.792956] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 744.794030] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 744.794512] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 744.794772] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 744.794912] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 744.795155] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 744.795326] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 744.795504] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 744.795660] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 744.796100] env[65522]: DEBUG nova.virt.hardware [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 744.804218] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Reconfiguring VM instance instance-00000019 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 744.805306] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5b2a557-0f69-4755-b76c-c31e4cbe3ea0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.830161] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 744.830161] env[65522]: value = "task-5113853" [ 744.830161] env[65522]: _type = "Task" [ 744.830161] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.831337] env[65522]: WARNING openstack [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.833473] env[65522]: WARNING openstack [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.855131] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113853, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.910310] env[65522]: DEBUG nova.compute.manager [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Received event network-vif-deleted-bb734aa2-6103-4321-b1bd-f9f5a80296ed {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 744.911491] env[65522]: DEBUG nova.compute.manager [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Received event network-vif-plugged-93f5a0f6-2499-4475-b6e9-68d105b143f3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 744.913016] env[65522]: DEBUG oslo_concurrency.lockutils [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Acquiring lock "028c7fc5-5624-4dd5-9e2f-48191a86f765-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 744.913016] env[65522]: DEBUG oslo_concurrency.lockutils [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Lock "028c7fc5-5624-4dd5-9e2f-48191a86f765-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 744.913016] env[65522]: DEBUG oslo_concurrency.lockutils [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Lock "028c7fc5-5624-4dd5-9e2f-48191a86f765-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.913016] env[65522]: DEBUG nova.compute.manager [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] No waiting events found dispatching network-vif-plugged-93f5a0f6-2499-4475-b6e9-68d105b143f3 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 744.913016] env[65522]: WARNING nova.compute.manager [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Received unexpected event network-vif-plugged-93f5a0f6-2499-4475-b6e9-68d105b143f3 for instance with vm_state building and task_state spawning. [ 744.913016] env[65522]: DEBUG nova.compute.manager [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Received event network-changed-93f5a0f6-2499-4475-b6e9-68d105b143f3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 744.913016] env[65522]: DEBUG nova.compute.manager [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Refreshing instance network info cache due to event network-changed-93f5a0f6-2499-4475-b6e9-68d105b143f3. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 744.913366] env[65522]: DEBUG oslo_concurrency.lockutils [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Acquiring lock "refresh_cache-028c7fc5-5624-4dd5-9e2f-48191a86f765" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.913366] env[65522]: DEBUG oslo_concurrency.lockutils [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Acquired lock "refresh_cache-028c7fc5-5624-4dd5-9e2f-48191a86f765" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 744.913454] env[65522]: DEBUG nova.network.neutron [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Refreshing network info cache for port 93f5a0f6-2499-4475-b6e9-68d105b143f3 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 744.918178] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 744.942241] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113850, 'name': Rename_Task, 'duration_secs': 0.19583} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.951584] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 744.955550] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-112ce002-ef00-4866-be22-3c534cdd23a6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.969090] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 744.969090] env[65522]: value = "task-5113854" [ 744.969090] env[65522]: _type = "Task" [ 744.969090] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.970662] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 744.976650] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac7a59e8-9296-4146-a2aa-6491b56433b2 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "6c8165f5-6769-4954-b9ac-de8551954377" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.504s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.986942] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113854, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.009564] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.010261] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.020844] env[65522]: INFO nova.compute.manager [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Rebuilding instance [ 745.079148] env[65522]: WARNING neutronclient.v2_0.client [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.080672] env[65522]: WARNING openstack [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.081302] env[65522]: WARNING openstack [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.128482] env[65522]: DEBUG nova.compute.manager [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 745.128992] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113851, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.130822] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c346d2d-ee53-4aa9-81cd-c97257f1f684 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.150158] env[65522]: WARNING neutronclient.v2_0.client [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.151030] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.151413] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.177913] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113852, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.286279] env[65522]: DEBUG nova.network.neutron [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Updating instance_info_cache with network_info: [{"id": "345f682f-4aaa-4bc1-bb61-e9d25e6288c5", "address": "fa:16:3e:d4:18:85", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.243", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap345f682f-4a", "ovs_interfaceid": "345f682f-4aaa-4bc1-bb61-e9d25e6288c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 745.354892] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113853, 'name': ReconfigVM_Task, 'duration_secs': 0.262498} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.358344] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Reconfigured VM instance instance-00000019 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 745.360941] env[65522]: DEBUG nova.network.neutron [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Updated VIF entry in instance network info cache for port 847d52c1-cfc5-4051-9969-5b8b567f330b. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 745.361312] env[65522]: DEBUG nova.network.neutron [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Updating instance_info_cache with network_info: [{"id": "847d52c1-cfc5-4051-9969-5b8b567f330b", "address": "fa:16:3e:c6:f3:5e", "network": {"id": "f83e6c69-de0c-4306-ac88-4ad211e44d55", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1504496554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "29c2f8cc328d4f6b985d258442c440b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "392517be-6cb8-4b5b-9a52-449bfe2e16f7", "external-id": "nsx-vlan-transportzone-351", "segmentation_id": 351, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap847d52c1-cf", "ovs_interfaceid": "847d52c1-cfc5-4051-9969-5b8b567f330b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 745.363330] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fa2d85-9526-4ea8-ae7e-83a5bcf5a865 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.395961] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Reconfiguring VM instance instance-00000019 to attach disk [datastore1] d5e8dd05-dc3c-4831-b4b0-ac100360f3e7/d5e8dd05-dc3c-4831-b4b0-ac100360f3e7.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 745.402084] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8ca32d3-76b9-4610-b049-bff74eb39eef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.423633] env[65522]: WARNING neutronclient.v2_0.client [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.424358] env[65522]: WARNING openstack [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.424775] env[65522]: WARNING openstack [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.442155] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 745.442155] env[65522]: value = "task-5113855" [ 745.442155] env[65522]: _type = "Task" [ 745.442155] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.453361] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113855, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.490606] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113854, 'name': PowerOnVM_Task} progress is 79%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.576701] env[65522]: WARNING openstack [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.577364] env[65522]: WARNING openstack [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.593927] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113851, 'name': PowerOnVM_Task, 'duration_secs': 0.613921} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.594532] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 745.594852] env[65522]: INFO nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Took 15.01 seconds to spawn the instance on the hypervisor. [ 745.595108] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 745.596257] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20a0d87-9c33-4409-95f0-6d53ee21d653 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.647670] env[65522]: WARNING neutronclient.v2_0.client [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.648953] env[65522]: WARNING openstack [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.649573] env[65522]: WARNING openstack [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.679435] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113852, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.588516} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.683393] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 028c7fc5-5624-4dd5-9e2f-48191a86f765/028c7fc5-5624-4dd5-9e2f-48191a86f765.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 745.683780] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 745.685053] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c51cb222-7318-4899-a320-7fa83522c9b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.700129] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 745.700129] env[65522]: value = "task-5113856" [ 745.700129] env[65522]: _type = "Task" [ 745.700129] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.717095] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113856, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.794286] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Releasing lock "refresh_cache-da211708-a6a2-4e03-bb41-c5b93564de20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 745.794881] env[65522]: DEBUG nova.compute.manager [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Instance network_info: |[{"id": "345f682f-4aaa-4bc1-bb61-e9d25e6288c5", "address": "fa:16:3e:d4:18:85", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.243", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap345f682f-4a", "ovs_interfaceid": "345f682f-4aaa-4bc1-bb61-e9d25e6288c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 745.797447] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:18:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '345f682f-4aaa-4bc1-bb61-e9d25e6288c5', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 745.811048] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 745.811048] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 745.811048] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74e5c5ab-daf9-4b56-b65c-4bf9aea9e926 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.834626] env[65522]: DEBUG nova.network.neutron [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Updated VIF entry in instance network info cache for port 93f5a0f6-2499-4475-b6e9-68d105b143f3. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 745.835329] env[65522]: DEBUG nova.network.neutron [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Updating instance_info_cache with network_info: [{"id": "93f5a0f6-2499-4475-b6e9-68d105b143f3", "address": "fa:16:3e:f6:b8:ff", "network": {"id": "f83e6c69-de0c-4306-ac88-4ad211e44d55", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1504496554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "29c2f8cc328d4f6b985d258442c440b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "392517be-6cb8-4b5b-9a52-449bfe2e16f7", "external-id": "nsx-vlan-transportzone-351", "segmentation_id": 351, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93f5a0f6-24", "ovs_interfaceid": "93f5a0f6-2499-4475-b6e9-68d105b143f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 745.842558] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d70574-ab3c-4096-ab95-119db44a3ac2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.853431] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3604aebf-3f5f-4d89-a931-13dedda53ebc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.859633] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 745.859633] env[65522]: value = "task-5113857" [ 745.859633] env[65522]: _type = "Task" [ 745.859633] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.896849] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Releasing lock "refresh_cache-428873b8-1e58-4924-86f3-8feca8f53342" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 745.896920] env[65522]: DEBUG nova.compute.manager [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Received event network-changed-37951fc0-2513-48ca-8e6a-42be7de65465 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 745.897088] env[65522]: DEBUG nova.compute.manager [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Refreshing instance network info cache due to event network-changed-37951fc0-2513-48ca-8e6a-42be7de65465. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 745.897311] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Acquiring lock "refresh_cache-75c8848b-aea9-43f1-8697-9224050d1fef" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.897492] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Acquired lock "refresh_cache-75c8848b-aea9-43f1-8697-9224050d1fef" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 745.897664] env[65522]: DEBUG nova.network.neutron [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Refreshing network info cache for port 37951fc0-2513-48ca-8e6a-42be7de65465 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 745.905513] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c35f5285-3827-4491-81e7-f8b89b0b378a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.908670] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113857, 'name': CreateVM_Task} progress is 15%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.917557] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b335153-7ee8-435b-81c8-8bea1366b296 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.924811] env[65522]: INFO nova.compute.manager [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Unrescuing [ 745.926649] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "refresh_cache-ca964440-5375-4aff-8b45-96fbe829dd16" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.926649] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "refresh_cache-ca964440-5375-4aff-8b45-96fbe829dd16" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 745.926649] env[65522]: DEBUG nova.network.neutron [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 745.941948] env[65522]: DEBUG nova.compute.provider_tree [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.954924] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113855, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.985166] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113854, 'name': PowerOnVM_Task, 'duration_secs': 0.845105} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.989722] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 745.989722] env[65522]: INFO nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Took 12.58 seconds to spawn the instance on the hypervisor. [ 745.989722] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 745.990684] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26120e5-e37f-43ce-96ce-0ec1e58559e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.032331] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7631ecd8-c3ea-4ddf-bc11-2ebf867cf820 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 746.032575] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7631ecd8-c3ea-4ddf-bc11-2ebf867cf820 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 746.033960] env[65522]: DEBUG nova.compute.manager [None req-7631ecd8-c3ea-4ddf-bc11-2ebf867cf820 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 746.033960] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "bb20d99f-fc49-47bf-a6b7-74ea966b71eb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 746.033960] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "bb20d99f-fc49-47bf-a6b7-74ea966b71eb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 746.035595] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f402f48a-27e1-4d6d-bdda-b6884d9ec682 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.045202] env[65522]: DEBUG nova.compute.manager [None req-7631ecd8-c3ea-4ddf-bc11-2ebf867cf820 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65522) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 746.045921] env[65522]: DEBUG nova.objects.instance [None req-7631ecd8-c3ea-4ddf-bc11-2ebf867cf820 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lazy-loading 'flavor' on Instance uuid e5d2cc50-4ee6-4e64-9b52-888968a717ca {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 746.073377] env[65522]: DEBUG nova.compute.manager [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Received event network-vif-plugged-95b7c754-4f56-4bea-84f7-f4269d730b51 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 746.075327] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Acquiring lock "54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 746.075327] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Lock "54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 746.075327] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Lock "54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 746.075327] env[65522]: DEBUG nova.compute.manager [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] No waiting events found dispatching network-vif-plugged-95b7c754-4f56-4bea-84f7-f4269d730b51 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 746.075327] env[65522]: WARNING nova.compute.manager [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Received unexpected event network-vif-plugged-95b7c754-4f56-4bea-84f7-f4269d730b51 for instance with vm_state building and task_state spawning. [ 746.075327] env[65522]: DEBUG nova.compute.manager [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Received event network-changed-95b7c754-4f56-4bea-84f7-f4269d730b51 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 746.075327] env[65522]: DEBUG nova.compute.manager [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Refreshing instance network info cache due to event network-changed-95b7c754-4f56-4bea-84f7-f4269d730b51. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 746.075327] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Acquiring lock "refresh_cache-54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.075327] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Acquired lock "refresh_cache-54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 746.075918] env[65522]: DEBUG nova.network.neutron [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Refreshing network info cache for port 95b7c754-4f56-4bea-84f7-f4269d730b51 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 746.076999] env[65522]: WARNING oslo_messaging._drivers.amqpdriver [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 746.123008] env[65522]: INFO nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Took 26.64 seconds to build instance. [ 746.137732] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquiring lock "6abf58ca-ad6b-4fe8-844a-ca33e62a844d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 746.138568] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Lock "6abf58ca-ad6b-4fe8-844a-ca33e62a844d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 746.138568] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquiring lock "6abf58ca-ad6b-4fe8-844a-ca33e62a844d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 746.138996] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Lock "6abf58ca-ad6b-4fe8-844a-ca33e62a844d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 746.139271] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Lock "6abf58ca-ad6b-4fe8-844a-ca33e62a844d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 746.142833] env[65522]: INFO nova.compute.manager [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Terminating instance [ 746.173677] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 746.173677] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c01d1d17-788d-4ec7-a5a6-9162a1231d1a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.186196] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Waiting for the task: (returnval){ [ 746.186196] env[65522]: value = "task-5113858" [ 746.186196] env[65522]: _type = "Task" [ 746.186196] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.197828] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113858, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.214319] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113856, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102608} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.215150] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 746.215626] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a46d0cb-db12-476c-9a9d-a3ba65918592 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.240294] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Reconfiguring VM instance instance-00000024 to attach disk [datastore1] 028c7fc5-5624-4dd5-9e2f-48191a86f765/028c7fc5-5624-4dd5-9e2f-48191a86f765.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 746.240508] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43d85d53-f394-4eb8-8f86-507d64fb2e04 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.262088] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 746.262088] env[65522]: value = "task-5113859" [ 746.262088] env[65522]: _type = "Task" [ 746.262088] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.278027] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113859, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.346598] env[65522]: DEBUG oslo_concurrency.lockutils [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Releasing lock "refresh_cache-028c7fc5-5624-4dd5-9e2f-48191a86f765" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 746.346913] env[65522]: DEBUG nova.compute.manager [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Received event network-changed-37951fc0-2513-48ca-8e6a-42be7de65465 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 746.347057] env[65522]: DEBUG nova.compute.manager [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Refreshing instance network info cache due to event network-changed-37951fc0-2513-48ca-8e6a-42be7de65465. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 746.347245] env[65522]: DEBUG oslo_concurrency.lockutils [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Acquiring lock "refresh_cache-75c8848b-aea9-43f1-8697-9224050d1fef" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.373913] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113857, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.405591] env[65522]: WARNING neutronclient.v2_0.client [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 746.406716] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 746.407030] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 746.433620] env[65522]: WARNING neutronclient.v2_0.client [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 746.435058] env[65522]: WARNING openstack [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 746.436196] env[65522]: WARNING openstack [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 746.448429] env[65522]: DEBUG nova.scheduler.client.report [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 746.460365] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113855, 'name': ReconfigVM_Task, 'duration_secs': 0.670391} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.460755] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Reconfigured VM instance instance-00000019 to attach disk [datastore1] d5e8dd05-dc3c-4831-b4b0-ac100360f3e7/d5e8dd05-dc3c-4831-b4b0-ac100360f3e7.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 746.461167] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating instance 'd5e8dd05-dc3c-4831-b4b0-ac100360f3e7' progress to 50 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 746.509859] env[65522]: INFO nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Took 26.96 seconds to build instance. [ 746.540389] env[65522]: DEBUG nova.compute.manager [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 746.544273] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 746.544638] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 746.580583] env[65522]: WARNING neutronclient.v2_0.client [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 746.581320] env[65522]: WARNING openstack [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 746.581677] env[65522]: WARNING openstack [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 746.590535] env[65522]: WARNING openstack [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 746.590911] env[65522]: WARNING openstack [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 746.625901] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "428873b8-1e58-4924-86f3-8feca8f53342" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.160s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 746.648571] env[65522]: WARNING neutronclient.v2_0.client [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 746.649349] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 746.649744] env[65522]: WARNING openstack [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 746.658212] env[65522]: DEBUG nova.compute.manager [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 746.658829] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 746.659638] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e51f2ebd-3c0e-4463-bb55-f7e507fca797 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.671140] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 746.671550] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e7d5997-eaf7-4962-aa62-3c4399dc3196 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.683997] env[65522]: DEBUG oslo_vmware.api [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 746.683997] env[65522]: value = "task-5113860" [ 746.683997] env[65522]: _type = "Task" [ 746.683997] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.705714] env[65522]: DEBUG oslo_vmware.api [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113860, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.708966] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113858, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.735171] env[65522]: WARNING neutronclient.v2_0.client [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 746.736234] env[65522]: WARNING openstack [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 746.737185] env[65522]: WARNING openstack [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 746.763065] env[65522]: WARNING openstack [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 746.763524] env[65522]: WARNING openstack [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 746.783249] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113859, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.808292] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 746.808666] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 746.809101] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 746.809320] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 746.810493] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 746.813224] env[65522]: INFO nova.compute.manager [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Terminating instance [ 746.868766] env[65522]: DEBUG nova.network.neutron [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Updated VIF entry in instance network info cache for port 37951fc0-2513-48ca-8e6a-42be7de65465. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 746.869250] env[65522]: DEBUG nova.network.neutron [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Updating instance_info_cache with network_info: [{"id": "37951fc0-2513-48ca-8e6a-42be7de65465", "address": "fa:16:3e:5e:53:01", "network": {"id": "922eeeeb-e124-4ae5-97a4-0e995b16f7ec", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-437359683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37fc01876567476f9b93d765b2cfddc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccf76700-491b-4462-ab19-e6d3a9ff87ac", "external-id": "nsx-vlan-transportzone-956", "segmentation_id": 956, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37951fc0-25", "ovs_interfaceid": "37951fc0-2513-48ca-8e6a-42be7de65465", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 746.877806] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113857, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.894126] env[65522]: WARNING neutronclient.v2_0.client [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 746.894871] env[65522]: WARNING openstack [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 746.895273] env[65522]: WARNING openstack [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 746.930129] env[65522]: DEBUG nova.network.neutron [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Updating instance_info_cache with network_info: [{"id": "ede8edfa-842b-482d-8e7b-34bde13a02f8", "address": "fa:16:3e:f6:50:03", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapede8edfa-84", "ovs_interfaceid": "ede8edfa-842b-482d-8e7b-34bde13a02f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 746.955933] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.569s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 746.958158] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.576s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 746.958413] env[65522]: DEBUG nova.objects.instance [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lazy-loading 'resources' on Instance uuid 57d32fb8-885e-44ba-967b-69f4bd1f5744 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 746.971410] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22738eae-a91a-4dac-9be4-8d9c66c87ba6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.996946] env[65522]: INFO nova.scheduler.client.report [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Deleted allocations for instance 1d3a9ada-df46-46af-a7f1-a84c251af4ac [ 746.999057] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d524c7e1-893f-46d8-b2d4-5df92e2b821e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.026934] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.486s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 747.027727] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating instance 'd5e8dd05-dc3c-4831-b4b0-ac100360f3e7' progress to 67 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 747.036046] env[65522]: DEBUG nova.network.neutron [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Updated VIF entry in instance network info cache for port 95b7c754-4f56-4bea-84f7-f4269d730b51. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 747.036046] env[65522]: DEBUG nova.network.neutron [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Updating instance_info_cache with network_info: [{"id": "95b7c754-4f56-4bea-84f7-f4269d730b51", "address": "fa:16:3e:30:82:2b", "network": {"id": "f83e6c69-de0c-4306-ac88-4ad211e44d55", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1504496554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "29c2f8cc328d4f6b985d258442c440b8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "392517be-6cb8-4b5b-9a52-449bfe2e16f7", "external-id": "nsx-vlan-transportzone-351", "segmentation_id": 351, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b7c754-4f", "ovs_interfaceid": "95b7c754-4f56-4bea-84f7-f4269d730b51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 747.065378] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7631ecd8-c3ea-4ddf-bc11-2ebf867cf820 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 747.066065] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-810f0bed-d926-4910-bec4-290f0d367025 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.076354] env[65522]: DEBUG oslo_vmware.api [None req-7631ecd8-c3ea-4ddf-bc11-2ebf867cf820 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 747.076354] env[65522]: value = "task-5113861" [ 747.076354] env[65522]: _type = "Task" [ 747.076354] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.077961] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.088856] env[65522]: DEBUG oslo_vmware.api [None req-7631ecd8-c3ea-4ddf-bc11-2ebf867cf820 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113861, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.199576] env[65522]: DEBUG oslo_vmware.api [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113860, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.207555] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113858, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.282725] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113859, 'name': ReconfigVM_Task, 'duration_secs': 0.840056} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.283042] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Reconfigured VM instance instance-00000024 to attach disk [datastore1] 028c7fc5-5624-4dd5-9e2f-48191a86f765/028c7fc5-5624-4dd5-9e2f-48191a86f765.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 747.283834] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0bb39190-ca5d-42c3-b3e7-ce65e13b1ddc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.293200] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 747.293200] env[65522]: value = "task-5113862" [ 747.293200] env[65522]: _type = "Task" [ 747.293200] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.307994] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113862, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.318092] env[65522]: DEBUG nova.compute.manager [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 747.318417] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 747.319530] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a766cb82-1758-4387-a0dc-e47be6e66ab9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.329358] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 747.329718] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1cf9c1dd-4700-42a3-8460-3fc879854e72 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.339121] env[65522]: DEBUG oslo_vmware.api [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 747.339121] env[65522]: value = "task-5113863" [ 747.339121] env[65522]: _type = "Task" [ 747.339121] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.351574] env[65522]: DEBUG oslo_vmware.api [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113863, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.376993] env[65522]: DEBUG oslo_concurrency.lockutils [req-84e19072-6cc1-4e87-a69c-32636539836f req-38316d66-4b77-470e-9eed-6e1cccae3d3c service nova] Releasing lock "refresh_cache-75c8848b-aea9-43f1-8697-9224050d1fef" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 747.377906] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113857, 'name': CreateVM_Task, 'duration_secs': 1.464234} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.378261] env[65522]: DEBUG oslo_concurrency.lockutils [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Acquired lock "refresh_cache-75c8848b-aea9-43f1-8697-9224050d1fef" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 747.378400] env[65522]: DEBUG nova.network.neutron [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Refreshing network info cache for port 37951fc0-2513-48ca-8e6a-42be7de65465 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 747.380661] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 747.381653] env[65522]: WARNING neutronclient.v2_0.client [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 747.382286] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.382510] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 747.382837] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 747.383278] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6faa7292-b126-4429-98e8-917dc8253ff2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.393031] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 747.393031] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527da7c9-89fe-177e-f3ba-26ee4306045d" [ 747.393031] env[65522]: _type = "Task" [ 747.393031] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.408956] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527da7c9-89fe-177e-f3ba-26ee4306045d, 'name': SearchDatastore_Task, 'duration_secs': 0.013118} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.409559] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 747.409559] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 747.409851] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.410058] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 747.410341] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 747.410757] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01efa376-f8be-44b0-b582-f818f683c2ab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.428862] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 747.429158] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 747.430339] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47775296-9f24-4c92-b6fe-2e650c28c96f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.435077] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "refresh_cache-ca964440-5375-4aff-8b45-96fbe829dd16" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 747.435353] env[65522]: DEBUG nova.objects.instance [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lazy-loading 'flavor' on Instance uuid ca964440-5375-4aff-8b45-96fbe829dd16 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 747.443728] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 747.443728] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524832d5-17f1-55ab-30b9-79c7e00042c2" [ 747.443728] env[65522]: _type = "Task" [ 747.443728] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.457566] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524832d5-17f1-55ab-30b9-79c7e00042c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.512835] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e5f054d-1de4-4ecc-8d27-09595ce3db62 tempest-ImagesOneServerTestJSON-419256281 tempest-ImagesOneServerTestJSON-419256281-project-member] Lock "1d3a9ada-df46-46af-a7f1-a84c251af4ac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.013s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 747.541957] env[65522]: WARNING neutronclient.v2_0.client [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 747.544926] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Releasing lock "refresh_cache-54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 747.545891] env[65522]: DEBUG nova.compute.manager [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Received event network-vif-plugged-345f682f-4aaa-4bc1-bb61-e9d25e6288c5 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 747.545891] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Acquiring lock "da211708-a6a2-4e03-bb41-c5b93564de20-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.545891] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Lock "da211708-a6a2-4e03-bb41-c5b93564de20-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.545891] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] Lock "da211708-a6a2-4e03-bb41-c5b93564de20-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 747.546356] env[65522]: DEBUG nova.compute.manager [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] No waiting events found dispatching network-vif-plugged-345f682f-4aaa-4bc1-bb61-e9d25e6288c5 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 747.546356] env[65522]: WARNING nova.compute.manager [req-5f00379c-6b8a-4781-b947-012c79776395 req-88dfb3d4-fbbc-4e2d-b52a-b449150d9f25 service nova] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Received unexpected event network-vif-plugged-345f682f-4aaa-4bc1-bb61-e9d25e6288c5 for instance with vm_state building and task_state spawning. [ 747.592033] env[65522]: DEBUG oslo_vmware.api [None req-7631ecd8-c3ea-4ddf-bc11-2ebf867cf820 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113861, 'name': PowerOffVM_Task, 'duration_secs': 0.358943} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.592350] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7631ecd8-c3ea-4ddf-bc11-2ebf867cf820 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 747.592546] env[65522]: DEBUG nova.compute.manager [None req-7631ecd8-c3ea-4ddf-bc11-2ebf867cf820 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 747.593394] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b454d40c-ecfd-43d9-890b-299866dd0fbd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.624949] env[65522]: DEBUG nova.network.neutron [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Port fb1cd05a-757b-469c-9749-fd1ca4fcb168 binding to destination host cpu-1 is already ACTIVE {{(pid=65522) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 747.709983] env[65522]: DEBUG oslo_vmware.api [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113860, 'name': PowerOffVM_Task, 'duration_secs': 0.75677} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.717062] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 747.717209] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 747.717519] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113858, 'name': PowerOffVM_Task, 'duration_secs': 1.20887} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.718198] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e15dfba-42d8-4f9f-a8e2-66f3a2ac65a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.722335] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 747.722335] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 747.722866] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-958ef9b1-1872-4540-9d9b-e443963e9cbc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.739194] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 747.742433] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07e035f4-fb7f-4a22-9ff7-add496c8b88b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.775392] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 747.775515] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 747.775899] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Deleting the datastore file [datastore1] 6e10e846-cdb7-458e-b511-18ab1742228e {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 747.776279] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0392c363-6446-4135-8c92-3578182fca91 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.785409] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Waiting for the task: (returnval){ [ 747.785409] env[65522]: value = "task-5113866" [ 747.785409] env[65522]: _type = "Task" [ 747.785409] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.802418] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113866, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.805456] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 747.805689] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 747.806019] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Deleting the datastore file [datastore2] 6abf58ca-ad6b-4fe8-844a-ca33e62a844d {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 747.806222] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-31fd1cfa-0a72-4ccd-9adb-8b534c208c8a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.812919] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113862, 'name': Rename_Task, 'duration_secs': 0.190808} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.813873] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 747.814221] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1cb7ed46-906e-43d3-a28c-df0a0283155a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.819808] env[65522]: DEBUG oslo_vmware.api [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for the task: (returnval){ [ 747.819808] env[65522]: value = "task-5113867" [ 747.819808] env[65522]: _type = "Task" [ 747.819808] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.829073] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 747.829073] env[65522]: value = "task-5113868" [ 747.829073] env[65522]: _type = "Task" [ 747.829073] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.836931] env[65522]: DEBUG oslo_vmware.api [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113867, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.847377] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113868, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.860191] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "24f520ec-6a56-4f17-9ae4-2c856bf99582" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.861290] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "24f520ec-6a56-4f17-9ae4-2c856bf99582" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.862438] env[65522]: DEBUG oslo_vmware.api [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113863, 'name': PowerOffVM_Task, 'duration_secs': 0.27117} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.863434] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 747.863712] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 747.864100] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Acquiring lock "20855931-a02d-4b53-b998-3d89d4ff5d1a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.864735] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Lock "20855931-a02d-4b53-b998-3d89d4ff5d1a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.864735] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Acquiring lock "20855931-a02d-4b53-b998-3d89d4ff5d1a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.864942] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Lock "20855931-a02d-4b53-b998-3d89d4ff5d1a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.865170] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Lock "20855931-a02d-4b53-b998-3d89d4ff5d1a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 747.867430] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7b3043c-5cfe-43d7-ad6f-d40c3835cd85 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.870332] env[65522]: INFO nova.compute.manager [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Terminating instance [ 747.882453] env[65522]: WARNING neutronclient.v2_0.client [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 747.883971] env[65522]: WARNING openstack [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 747.884463] env[65522]: WARNING openstack [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 747.942698] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90fb07e3-e322-4e08-9363-aae8acdbd85f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.964285] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524832d5-17f1-55ab-30b9-79c7e00042c2, 'name': SearchDatastore_Task, 'duration_secs': 0.013064} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.985753] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 747.990854] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db64d492-e9f9-4000-9a0a-10e610de7144 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.993366] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58673f33-8a35-4a59-b0f7-2de204113eef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.995057] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 747.995900] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 747.995900] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleting the datastore file [datastore2] f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 747.996312] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3fb8fe86-c40d-440a-9196-039ce9e0a99a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.003443] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 748.003443] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5218b8cc-76d3-f5f7-341f-4637eb5db90e" [ 748.003443] env[65522]: _type = "Task" [ 748.003443] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.013507] env[65522]: DEBUG oslo_vmware.api [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 748.013507] env[65522]: value = "task-5113871" [ 748.013507] env[65522]: _type = "Task" [ 748.013507] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.013507] env[65522]: DEBUG oslo_vmware.api [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 748.013507] env[65522]: value = "task-5113870" [ 748.013507] env[65522]: _type = "Task" [ 748.013507] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.024306] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5218b8cc-76d3-f5f7-341f-4637eb5db90e, 'name': SearchDatastore_Task} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.029274] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 748.029274] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] da211708-a6a2-4e03-bb41-c5b93564de20/da211708-a6a2-4e03-bb41-c5b93564de20.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 748.029274] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54f60862-858c-4387-86a7-5d4c451e2c77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.036089] env[65522]: DEBUG oslo_vmware.api [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113871, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.036352] env[65522]: DEBUG oslo_vmware.api [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113870, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.044690] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 748.044690] env[65522]: value = "task-5113872" [ 748.044690] env[65522]: _type = "Task" [ 748.044690] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.055945] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113872, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.105495] env[65522]: WARNING openstack [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.105495] env[65522]: WARNING openstack [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.120872] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7631ecd8-c3ea-4ddf-bc11-2ebf867cf820 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.088s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 748.243752] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b13d45c-5e55-4e60-ad95-59ccdb088edf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.252570] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc98dd7b-f196-4322-85b3-8d6a635de4e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.287800] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97968d88-458b-48c4-8599-7459afeae465 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.305423] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd7eb6b-0ad1-408c-b761-03a62f16d928 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.309207] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113866, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129462} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.309533] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 748.309897] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 748.310081] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.329228] env[65522]: DEBUG nova.compute.provider_tree [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.348816] env[65522]: DEBUG oslo_vmware.api [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Task: {'id': task-5113867, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170053} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.353429] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 748.354030] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 748.354030] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.354127] env[65522]: INFO nova.compute.manager [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Took 1.70 seconds to destroy the instance on the hypervisor. [ 748.354368] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 748.354677] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113868, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.355821] env[65522]: DEBUG nova.compute.manager [-] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 748.355821] env[65522]: DEBUG nova.network.neutron [-] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 748.355943] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.356666] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.356989] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.366374] env[65522]: DEBUG nova.compute.manager [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 748.378900] env[65522]: DEBUG nova.compute.manager [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 748.379336] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 748.382092] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487b2f5b-d32b-46ad-b4ca-405d4a66faf0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.395101] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 748.395436] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a4f9d4e-81be-47b6-9985-0041eaf2095f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.405277] env[65522]: DEBUG oslo_vmware.api [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Waiting for the task: (returnval){ [ 748.405277] env[65522]: value = "task-5113873" [ 748.405277] env[65522]: _type = "Task" [ 748.405277] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.416996] env[65522]: DEBUG oslo_vmware.api [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113873, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.527769] env[65522]: DEBUG oslo_vmware.api [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113870, 'name': PowerOffVM_Task, 'duration_secs': 0.24404} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.532994] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 748.542673] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Reconfiguring VM instance instance-00000014 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 748.542673] env[65522]: DEBUG oslo_vmware.api [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113871, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183586} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.542673] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84ffbf2e-4b72-48b8-8739-b693eead5b01 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.561284] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 748.561284] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 748.561284] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.561284] env[65522]: INFO nova.compute.manager [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Took 1.24 seconds to destroy the instance on the hypervisor. [ 748.561284] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 748.561284] env[65522]: DEBUG nova.compute.manager [-] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 748.561284] env[65522]: DEBUG nova.network.neutron [-] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 748.561284] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.561748] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.561900] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.583171] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113872, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.585137] env[65522]: DEBUG oslo_vmware.api [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 748.585137] env[65522]: value = "task-5113874" [ 748.585137] env[65522]: _type = "Task" [ 748.585137] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.598062] env[65522]: DEBUG oslo_vmware.api [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113874, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.607180] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "495fa98f-a9db-4214-87cc-d29209d3cb62" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 748.607431] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "495fa98f-a9db-4214-87cc-d29209d3cb62" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 748.630201] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.653263] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 748.653654] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 748.653787] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 748.713549] env[65522]: WARNING neutronclient.v2_0.client [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.714447] env[65522]: WARNING openstack [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.714885] env[65522]: WARNING openstack [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.803444] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.840334] env[65522]: DEBUG nova.scheduler.client.report [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 748.855885] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113868, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.899095] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 748.921027] env[65522]: DEBUG oslo_vmware.api [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113873, 'name': PowerOffVM_Task, 'duration_secs': 0.474042} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.921027] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 748.921027] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 748.921027] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-81c78e17-ccfa-4796-bf77-e8ef8d1ee3cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.043125] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 749.043198] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 749.043404] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Deleting the datastore file [datastore2] 20855931-a02d-4b53-b998-3d89d4ff5d1a {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 749.043848] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-25e3fe1d-b579-4d33-8664-0aad94db1841 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.055715] env[65522]: DEBUG oslo_vmware.api [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Waiting for the task: (returnval){ [ 749.055715] env[65522]: value = "task-5113876" [ 749.055715] env[65522]: _type = "Task" [ 749.055715] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.075972] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113872, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637041} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.079500] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] da211708-a6a2-4e03-bb41-c5b93564de20/da211708-a6a2-4e03-bb41-c5b93564de20.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 749.079894] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 749.080353] env[65522]: DEBUG oslo_vmware.api [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113876, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.080624] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2984c678-6fd5-4a71-a79f-90ecc8fa0562 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.093059] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 749.093059] env[65522]: value = "task-5113877" [ 749.093059] env[65522]: _type = "Task" [ 749.093059] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.103842] env[65522]: DEBUG oslo_vmware.api [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113874, 'name': ReconfigVM_Task, 'duration_secs': 0.451714} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.107090] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Reconfigured VM instance instance-00000014 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 749.107090] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 749.107090] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2023e8ef-6079-42de-8381-639886a119dd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.113100] env[65522]: DEBUG nova.compute.manager [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 749.117565] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113877, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.128299] env[65522]: DEBUG oslo_vmware.api [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 749.128299] env[65522]: value = "task-5113878" [ 749.128299] env[65522]: _type = "Task" [ 749.128299] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.144729] env[65522]: DEBUG oslo_vmware.api [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113878, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.355409] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.397s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 749.358755] env[65522]: DEBUG oslo_vmware.api [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113868, 'name': PowerOnVM_Task, 'duration_secs': 1.194996} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.361102] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.385s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 749.361318] env[65522]: DEBUG nova.objects.instance [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 749.363999] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 749.364219] env[65522]: INFO nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Took 13.66 seconds to spawn the instance on the hypervisor. [ 749.364696] env[65522]: DEBUG nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 749.366117] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1a03b7-dc48-4a79-8d68-985dfc1be321 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.380663] env[65522]: DEBUG nova.virt.hardware [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 749.380893] env[65522]: DEBUG nova.virt.hardware [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 749.381087] env[65522]: DEBUG nova.virt.hardware [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 749.381248] env[65522]: DEBUG nova.virt.hardware [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 749.381386] env[65522]: DEBUG nova.virt.hardware [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 749.381525] env[65522]: DEBUG nova.virt.hardware [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 749.381737] env[65522]: DEBUG nova.virt.hardware [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.381882] env[65522]: DEBUG nova.virt.hardware [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 749.382059] env[65522]: DEBUG nova.virt.hardware [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 749.382222] env[65522]: DEBUG nova.virt.hardware [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 749.382391] env[65522]: DEBUG nova.virt.hardware [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 749.383331] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dacde3e1-8104-4935-bae6-b50b444745e2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.393326] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f80ed024-4265-4d29-95d2-718567431f77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.400870] env[65522]: INFO nova.scheduler.client.report [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Deleted allocations for instance 57d32fb8-885e-44ba-967b-69f4bd1f5744 [ 749.412561] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 749.419408] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 749.423437] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 749.423437] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-761859ad-9b49-4bc2-bbde-28266e850b3e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.450830] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 749.450830] env[65522]: value = "task-5113879" [ 749.450830] env[65522]: _type = "Task" [ 749.450830] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.463691] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113879, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.571794] env[65522]: DEBUG oslo_vmware.api [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113876, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.605968] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113877, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.123588} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.606293] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 749.607216] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca705a0a-64ee-437a-ad51-df6af1bd7f3a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.640051] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] da211708-a6a2-4e03-bb41-c5b93564de20/da211708-a6a2-4e03-bb41-c5b93564de20.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 749.640830] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-640002a5-a79d-4cf7-87ef-8c24e487ae77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.662761] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 749.668097] env[65522]: WARNING neutronclient.v2_0.client [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 749.678858] env[65522]: DEBUG oslo_vmware.api [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113878, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.680692] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 749.680692] env[65522]: value = "task-5113880" [ 749.680692] env[65522]: _type = "Task" [ 749.680692] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.690473] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113880, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.717951] env[65522]: DEBUG nova.network.neutron [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Updated VIF entry in instance network info cache for port 37951fc0-2513-48ca-8e6a-42be7de65465. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 749.718501] env[65522]: DEBUG nova.network.neutron [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Updating instance_info_cache with network_info: [{"id": "37951fc0-2513-48ca-8e6a-42be7de65465", "address": "fa:16:3e:5e:53:01", "network": {"id": "922eeeeb-e124-4ae5-97a4-0e995b16f7ec", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-437359683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37fc01876567476f9b93d765b2cfddc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccf76700-491b-4462-ab19-e6d3a9ff87ac", "external-id": "nsx-vlan-transportzone-956", "segmentation_id": 956, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37951fc0-25", "ovs_interfaceid": "37951fc0-2513-48ca-8e6a-42be7de65465", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 749.900325] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.900533] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 749.900856] env[65522]: DEBUG nova.network.neutron [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 749.903012] env[65522]: INFO nova.compute.manager [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Took 30.30 seconds to build instance. [ 749.927648] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc83b0a8-7e36-4a6a-998a-01a5bf6e06e1 tempest-FloatingIPsAssociationTestJSON-1340626663 tempest-FloatingIPsAssociationTestJSON-1340626663-project-member] Lock "57d32fb8-885e-44ba-967b-69f4bd1f5744" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.314s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 749.971607] env[65522]: DEBUG nova.network.neutron [-] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 749.972823] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113879, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.072721] env[65522]: DEBUG oslo_vmware.api [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Task: {'id': task-5113876, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.52049} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.072721] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 750.072721] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 750.072721] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 750.072721] env[65522]: INFO nova.compute.manager [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Took 1.69 seconds to destroy the instance on the hypervisor. [ 750.072721] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 750.072721] env[65522]: DEBUG nova.compute.manager [-] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 750.072721] env[65522]: DEBUG nova.network.neutron [-] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 750.072721] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 750.073264] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.073526] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.082537] env[65522]: DEBUG nova.network.neutron [-] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 750.154438] env[65522]: DEBUG oslo_vmware.api [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113878, 'name': PowerOnVM_Task, 'duration_secs': 0.718607} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.157959] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 750.159079] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 750.159242] env[65522]: DEBUG nova.compute.manager [None req-5e6dc780-4312-4ad5-94b1-82199c82dfe2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 750.160602] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c62ead1-68c5-400b-826e-87d5669df375 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.193729] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113880, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.223561] env[65522]: DEBUG oslo_concurrency.lockutils [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Releasing lock "refresh_cache-75c8848b-aea9-43f1-8697-9224050d1fef" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 750.223850] env[65522]: DEBUG nova.compute.manager [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Received event network-changed-dcf081b2-7a3e-4efd-84fa-bbf035dfd690 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 750.224036] env[65522]: DEBUG nova.compute.manager [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Refreshing instance network info cache due to event network-changed-dcf081b2-7a3e-4efd-84fa-bbf035dfd690. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 750.226357] env[65522]: DEBUG oslo_concurrency.lockutils [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Acquiring lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.226537] env[65522]: DEBUG oslo_concurrency.lockutils [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Acquired lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 750.226711] env[65522]: DEBUG nova.network.neutron [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Refreshing network info cache for port dcf081b2-7a3e-4efd-84fa-bbf035dfd690 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 750.376283] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6a1a37e-7d8d-474e-9a0c-e3bbf23162e9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 750.377598] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 13.972s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 750.377789] env[65522]: DEBUG nova.objects.instance [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 750.408099] env[65522]: WARNING neutronclient.v2_0.client [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 750.408099] env[65522]: WARNING openstack [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.408353] env[65522]: WARNING openstack [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.418138] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69995bdf-284a-4836-aa21-1ae28c8dcc4e tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "028c7fc5-5624-4dd5-9e2f-48191a86f765" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.822s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 750.463328] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113879, 'name': CreateVM_Task, 'duration_secs': 0.71349} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.463504] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 750.463981] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.464158] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 750.464589] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 750.467123] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6b4165a-2272-46e4-a140-82d89c68888e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.474400] env[65522]: INFO nova.compute.manager [-] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Took 2.12 seconds to deallocate network for instance. [ 750.474745] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Waiting for the task: (returnval){ [ 750.474745] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a85c4f-536c-18f8-c8f2-effbe103a543" [ 750.474745] env[65522]: _type = "Task" [ 750.474745] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.489431] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a85c4f-536c-18f8-c8f2-effbe103a543, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.585141] env[65522]: INFO nova.compute.manager [-] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Took 2.02 seconds to deallocate network for instance. [ 750.614323] env[65522]: WARNING openstack [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.614471] env[65522]: WARNING openstack [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.660585] env[65522]: DEBUG nova.compute.manager [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Received event network-changed-345f682f-4aaa-4bc1-bb61-e9d25e6288c5 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 750.660770] env[65522]: DEBUG nova.compute.manager [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Refreshing instance network info cache due to event network-changed-345f682f-4aaa-4bc1-bb61-e9d25e6288c5. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 750.661819] env[65522]: DEBUG oslo_concurrency.lockutils [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] Acquiring lock "refresh_cache-da211708-a6a2-4e03-bb41-c5b93564de20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.661819] env[65522]: DEBUG oslo_concurrency.lockutils [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] Acquired lock "refresh_cache-da211708-a6a2-4e03-bb41-c5b93564de20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 750.661819] env[65522]: DEBUG nova.network.neutron [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Refreshing network info cache for port 345f682f-4aaa-4bc1-bb61-e9d25e6288c5 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 750.688325] env[65522]: WARNING neutronclient.v2_0.client [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 750.689187] env[65522]: WARNING openstack [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.689711] env[65522]: WARNING openstack [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.704817] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113880, 'name': ReconfigVM_Task, 'duration_secs': 0.646469} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.705170] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Reconfigured VM instance instance-00000025 to attach disk [datastore1] da211708-a6a2-4e03-bb41-c5b93564de20/da211708-a6a2-4e03-bb41-c5b93564de20.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 750.705868] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d48baabe-07d4-40ee-8309-6989c0976abb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.713974] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 750.713974] env[65522]: value = "task-5113881" [ 750.713974] env[65522]: _type = "Task" [ 750.713974] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.724485] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113881, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.730636] env[65522]: WARNING neutronclient.v2_0.client [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 750.731575] env[65522]: WARNING openstack [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.731719] env[65522]: WARNING openstack [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.786674] env[65522]: DEBUG nova.network.neutron [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 750.852605] env[65522]: DEBUG nova.network.neutron [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating instance_info_cache with network_info: [{"id": "fb1cd05a-757b-469c-9749-fd1ca4fcb168", "address": "fa:16:3e:b1:9d:ad", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb1cd05a-75", "ovs_interfaceid": "fb1cd05a-757b-469c-9749-fd1ca4fcb168", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 750.911240] env[65522]: DEBUG nova.network.neutron [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 750.927986] env[65522]: DEBUG nova.network.neutron [-] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 750.990484] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 750.992755] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a85c4f-536c-18f8-c8f2-effbe103a543, 'name': SearchDatastore_Task, 'duration_secs': 0.035784} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.993238] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 750.994648] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 750.994648] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.994648] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 750.994648] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 750.994824] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0551158d-5cbd-4dff-b261-47f642c52de8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.009911] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 751.010265] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 751.010987] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a058f3a8-9215-46de-aaf7-f249a5ea8065 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.021365] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Waiting for the task: (returnval){ [ 751.021365] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52285383-1e10-b037-6453-a51f181639bb" [ 751.021365] env[65522]: _type = "Task" [ 751.021365] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.035761] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52285383-1e10-b037-6453-a51f181639bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.099960] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.163929] env[65522]: WARNING neutronclient.v2_0.client [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 751.164667] env[65522]: WARNING openstack [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 751.165196] env[65522]: WARNING openstack [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 751.229628] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113881, 'name': Rename_Task, 'duration_secs': 0.195162} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.230074] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 751.230444] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-20329d18-9f67-40bd-8b38-f8652696b393 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.239785] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 751.239785] env[65522]: value = "task-5113882" [ 751.239785] env[65522]: _type = "Task" [ 751.239785] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.246025] env[65522]: DEBUG nova.compute.manager [req-a0cf196c-7873-46af-8e1a-4095ebbe4836 req-3c203093-48bd-4d1e-83db-d047b1af9eaa service nova] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Received event network-vif-deleted-189bde05-0b1d-4c6a-be88-2453345dd731 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 751.251676] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113882, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.359360] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Releasing lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 751.389010] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6177b8e-2d0e-49ea-9f41-f9222229a166 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.390173] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.726s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.390398] env[65522]: DEBUG nova.objects.instance [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lazy-loading 'resources' on Instance uuid c0c99008-1907-4d3a-b4f6-ee6e916e2c67 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 751.413870] env[65522]: DEBUG oslo_concurrency.lockutils [req-b34a962c-b250-4556-81a4-3e5a27f158a1 req-577bda38-e7ec-4bd5-9f63-cd333ac668ac service nova] Releasing lock "refresh_cache-6abf58ca-ad6b-4fe8-844a-ca33e62a844d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 751.431107] env[65522]: INFO nova.compute.manager [-] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Took 1.36 seconds to deallocate network for instance. [ 751.465813] env[65522]: WARNING openstack [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 751.466225] env[65522]: WARNING openstack [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 751.540764] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52285383-1e10-b037-6453-a51f181639bb, 'name': SearchDatastore_Task, 'duration_secs': 0.040896} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.541783] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abf9d071-8cce-4ea1-bf36-5734f670a136 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.548863] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Waiting for the task: (returnval){ [ 751.548863] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52621fc6-db10-44e0-aeea-7cd60bb3c3ab" [ 751.548863] env[65522]: _type = "Task" [ 751.548863] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.561913] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52621fc6-db10-44e0-aeea-7cd60bb3c3ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.758736] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113882, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.829563] env[65522]: WARNING neutronclient.v2_0.client [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 751.830740] env[65522]: WARNING openstack [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 751.830893] env[65522]: WARNING openstack [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 751.900067] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26285092-29b0-4fe6-9b45-93d190007eb8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.926321] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d14412-1754-43ce-8772-5f9878cbb769 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.943470] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.947988] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating instance 'd5e8dd05-dc3c-4831-b4b0-ac100360f3e7' progress to 83 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 751.987341] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "7693f017-69b3-4ec4-ae3f-7e65ed34a471" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.987589] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "7693f017-69b3-4ec4-ae3f-7e65ed34a471" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 752.064411] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52621fc6-db10-44e0-aeea-7cd60bb3c3ab, 'name': SearchDatastore_Task, 'duration_secs': 0.015746} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.068890] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 752.069334] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 6e10e846-cdb7-458e-b511-18ab1742228e/6e10e846-cdb7-458e-b511-18ab1742228e.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 752.070738] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-034185dd-0e26-413b-b8e5-f92d428877a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.081698] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Waiting for the task: (returnval){ [ 752.081698] env[65522]: value = "task-5113883" [ 752.081698] env[65522]: _type = "Task" [ 752.081698] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.097801] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113883, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.142117] env[65522]: DEBUG nova.network.neutron [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Updated VIF entry in instance network info cache for port 345f682f-4aaa-4bc1-bb61-e9d25e6288c5. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 752.142365] env[65522]: DEBUG nova.network.neutron [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Updating instance_info_cache with network_info: [{"id": "345f682f-4aaa-4bc1-bb61-e9d25e6288c5", "address": "fa:16:3e:d4:18:85", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.243", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap345f682f-4a", "ovs_interfaceid": "345f682f-4aaa-4bc1-bb61-e9d25e6288c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 752.258923] env[65522]: DEBUG oslo_vmware.api [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113882, 'name': PowerOnVM_Task, 'duration_secs': 0.727309} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.259249] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 752.259460] env[65522]: INFO nova.compute.manager [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Took 10.47 seconds to spawn the instance on the hypervisor. [ 752.259638] env[65522]: DEBUG nova.compute.manager [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 752.260558] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff3a61cc-a6d7-4ce3-ad9a-f309c05f1ae7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.367023] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Acquiring lock "a09c12be-0aaa-4934-a1b3-5af29feba8a6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 752.367023] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Lock "a09c12be-0aaa-4934-a1b3-5af29feba8a6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 752.369512] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Acquiring lock "a09c12be-0aaa-4934-a1b3-5af29feba8a6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 752.369512] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Lock "a09c12be-0aaa-4934-a1b3-5af29feba8a6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.002s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 752.369512] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Lock "a09c12be-0aaa-4934-a1b3-5af29feba8a6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 752.378580] env[65522]: INFO nova.compute.manager [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Terminating instance [ 752.463064] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 752.463586] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e58ae655-ba81-4359-b322-1c5f79b0024f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.474997] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 752.474997] env[65522]: value = "task-5113884" [ 752.474997] env[65522]: _type = "Task" [ 752.474997] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.488845] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113884, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.490500] env[65522]: DEBUG nova.compute.manager [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 752.601763] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113883, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.638746] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6bb179d-bde2-4c31-b231-2ce2b40c3889 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.646073] env[65522]: DEBUG oslo_concurrency.lockutils [req-ce7277d9-bd8a-44fb-8be3-b72889a23ee9 req-f0e1b3a3-099a-40e4-b8be-6b4f1cebd6b8 service nova] Releasing lock "refresh_cache-da211708-a6a2-4e03-bb41-c5b93564de20" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 752.651715] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abeabab9-ddb2-4ff1-9e35-6c062e263a39 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.695211] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf023d9-cc54-4b71-9077-ac52fcfefbcb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.707933] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2ae6a7-f81c-4a01-b9e2-b8ad1e490ca5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.725690] env[65522]: DEBUG nova.compute.provider_tree [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.788977] env[65522]: INFO nova.compute.manager [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Took 31.93 seconds to build instance. [ 752.890838] env[65522]: DEBUG nova.compute.manager [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 752.890838] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 752.890838] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b78a74-a6c1-4941-88c6-a328d0949130 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.905471] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 752.905964] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1624a3fe-a270-4c18-82e9-da83787cb2bf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.918891] env[65522]: DEBUG oslo_vmware.api [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Waiting for the task: (returnval){ [ 752.918891] env[65522]: value = "task-5113885" [ 752.918891] env[65522]: _type = "Task" [ 752.918891] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.933502] env[65522]: DEBUG oslo_vmware.api [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113885, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.001082] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113884, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.029053] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 753.096496] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113883, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.9794} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.096716] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 6e10e846-cdb7-458e-b511-18ab1742228e/6e10e846-cdb7-458e-b511-18ab1742228e.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 753.096962] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 753.097240] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd78a95c-ae9b-4049-b8b9-69c8c2eb1690 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.105786] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Waiting for the task: (returnval){ [ 753.105786] env[65522]: value = "task-5113886" [ 753.105786] env[65522]: _type = "Task" [ 753.105786] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.120777] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113886, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.230416] env[65522]: DEBUG nova.scheduler.client.report [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 753.292175] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954adebc-7e4d-45ef-af5f-3c4afa1b2565 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "da211708-a6a2-4e03-bb41-c5b93564de20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.452s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 753.433574] env[65522]: DEBUG oslo_vmware.api [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113885, 'name': PowerOffVM_Task, 'duration_secs': 0.432266} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.433967] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 753.434085] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 753.434309] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3afb82fc-290f-404f-909d-e4b9d89fcba0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.502752] env[65522]: DEBUG oslo_vmware.api [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5113884, 'name': PowerOnVM_Task, 'duration_secs': 0.605013} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.502752] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 753.502752] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b8880898-5501-4841-af44-1b146ae49b4d tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating instance 'd5e8dd05-dc3c-4831-b4b0-ac100360f3e7' progress to 100 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 753.523596] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 753.524832] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 753.524832] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Deleting the datastore file [datastore2] a09c12be-0aaa-4934-a1b3-5af29feba8a6 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 753.524832] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e1c64841-3a80-4f47-9721-fd4ecf0ccf5d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.535960] env[65522]: DEBUG oslo_vmware.api [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Waiting for the task: (returnval){ [ 753.535960] env[65522]: value = "task-5113888" [ 753.535960] env[65522]: _type = "Task" [ 753.535960] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.551647] env[65522]: DEBUG oslo_vmware.api [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113888, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.618666] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113886, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107935} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.619243] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 753.620022] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31237686-e858-43b8-b668-c8f3f63a7d10 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.646946] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Reconfiguring VM instance instance-0000001c to attach disk [datastore2] 6e10e846-cdb7-458e-b511-18ab1742228e/6e10e846-cdb7-458e-b511-18ab1742228e.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 753.647449] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6026aa47-8f9e-458f-b991-a62744da1728 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.672215] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Waiting for the task: (returnval){ [ 753.672215] env[65522]: value = "task-5113889" [ 753.672215] env[65522]: _type = "Task" [ 753.672215] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.683188] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113889, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.744363] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.354s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 753.746914] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.198s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 753.751801] env[65522]: INFO nova.compute.claims [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.792113] env[65522]: INFO nova.scheduler.client.report [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Deleted allocations for instance c0c99008-1907-4d3a-b4f6-ee6e916e2c67 [ 753.868260] env[65522]: DEBUG nova.compute.manager [req-95f4aff4-b799-44ca-a15d-5ae81c19081a req-562d0e85-61d5-41af-a18b-826456c57d4b service nova] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Received event network-vif-deleted-dcf081b2-7a3e-4efd-84fa-bbf035dfd690 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 753.983724] env[65522]: DEBUG nova.objects.instance [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lazy-loading 'flavor' on Instance uuid e5d2cc50-4ee6-4e64-9b52-888968a717ca {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 753.987398] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3021d99c-8605-4f06-aaaa-6e41a511fedc tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "7c6f8218-602d-44f3-8012-de5a96972785" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 753.987620] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3021d99c-8605-4f06-aaaa-6e41a511fedc tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "7c6f8218-602d-44f3-8012-de5a96972785" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 753.987620] env[65522]: DEBUG nova.compute.manager [None req-3021d99c-8605-4f06-aaaa-6e41a511fedc tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 753.988519] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372f9610-a994-4778-8933-0215d42b98ca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.999061] env[65522]: DEBUG nova.compute.manager [None req-3021d99c-8605-4f06-aaaa-6e41a511fedc tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65522) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 753.999606] env[65522]: DEBUG nova.objects.instance [None req-3021d99c-8605-4f06-aaaa-6e41a511fedc tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lazy-loading 'flavor' on Instance uuid 7c6f8218-602d-44f3-8012-de5a96972785 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 754.052177] env[65522]: DEBUG oslo_vmware.api [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Task: {'id': task-5113888, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.369928} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.052177] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 754.052177] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 754.052177] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 754.052177] env[65522]: INFO nova.compute.manager [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Took 1.16 seconds to destroy the instance on the hypervisor. [ 754.052177] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 754.052177] env[65522]: DEBUG nova.compute.manager [-] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 754.052177] env[65522]: DEBUG nova.network.neutron [-] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 754.052177] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 754.052177] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.052177] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.104277] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 754.126844] env[65522]: DEBUG nova.compute.manager [req-586ba9c2-d58b-424d-b27c-90d9268ed9a9 req-e2c462d5-4c45-4e98-946b-6c1598833e81 service nova] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Received event network-vif-deleted-b7e73ab8-d1b6-4e32-845a-fb308037ddf9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 754.185902] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113889, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.306935] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9683a3af-bfc3-4c90-8fe7-5f8317e963cc tempest-AttachInterfacesV270Test-1217898182 tempest-AttachInterfacesV270Test-1217898182-project-member] Lock "c0c99008-1907-4d3a-b4f6-ee6e916e2c67" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.656s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 754.492750] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "refresh_cache-e5d2cc50-4ee6-4e64-9b52-888968a717ca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.492750] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquired lock "refresh_cache-e5d2cc50-4ee6-4e64-9b52-888968a717ca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 754.492750] env[65522]: DEBUG nova.network.neutron [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 754.492750] env[65522]: DEBUG nova.objects.instance [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lazy-loading 'info_cache' on Instance uuid e5d2cc50-4ee6-4e64-9b52-888968a717ca {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 754.686513] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113889, 'name': ReconfigVM_Task, 'duration_secs': 0.716398} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.686513] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Reconfigured VM instance instance-0000001c to attach disk [datastore2] 6e10e846-cdb7-458e-b511-18ab1742228e/6e10e846-cdb7-458e-b511-18ab1742228e.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 754.686513] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0860d564-3137-4f1e-9470-691fae3503f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.695246] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Waiting for the task: (returnval){ [ 754.695246] env[65522]: value = "task-5113890" [ 754.695246] env[65522]: _type = "Task" [ 754.695246] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.707759] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113890, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.719998] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "428873b8-1e58-4924-86f3-8feca8f53342" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 754.720206] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "428873b8-1e58-4924-86f3-8feca8f53342" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 754.720415] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "428873b8-1e58-4924-86f3-8feca8f53342-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 754.720615] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "428873b8-1e58-4924-86f3-8feca8f53342-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 754.720837] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "428873b8-1e58-4924-86f3-8feca8f53342-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 754.723944] env[65522]: INFO nova.compute.manager [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Terminating instance [ 754.996581] env[65522]: DEBUG nova.objects.base [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 755.006883] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-3021d99c-8605-4f06-aaaa-6e41a511fedc tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 755.007403] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-951d06fa-c651-46ff-8b7c-6bc2409b20d6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.020639] env[65522]: DEBUG oslo_vmware.api [None req-3021d99c-8605-4f06-aaaa-6e41a511fedc tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 755.020639] env[65522]: value = "task-5113891" [ 755.020639] env[65522]: _type = "Task" [ 755.020639] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.049119] env[65522]: DEBUG oslo_vmware.api [None req-3021d99c-8605-4f06-aaaa-6e41a511fedc tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113891, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.156548] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "interface-b2d70982-54e7-459c-a0d8-48bf4b6e4345-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 755.156615] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-b2d70982-54e7-459c-a0d8-48bf4b6e4345-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 755.157391] env[65522]: DEBUG nova.objects.instance [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'flavor' on Instance uuid b2d70982-54e7-459c-a0d8-48bf4b6e4345 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 755.207202] env[65522]: DEBUG nova.network.neutron [-] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 755.216492] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113890, 'name': Rename_Task, 'duration_secs': 0.195948} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.217014] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 755.217951] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4ed00aa-f052-4179-8cff-6695c65e910f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.228202] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Waiting for the task: (returnval){ [ 755.228202] env[65522]: value = "task-5113892" [ 755.228202] env[65522]: _type = "Task" [ 755.228202] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.235216] env[65522]: DEBUG nova.compute.manager [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 755.235216] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 755.236653] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64550b5c-4781-4952-aab2-252a4f7208ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.253027] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113892, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.253876] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 755.254189] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc8aede9-691a-4890-a6a1-8915ba8f7c66 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.263865] env[65522]: DEBUG oslo_vmware.api [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 755.263865] env[65522]: value = "task-5113893" [ 755.263865] env[65522]: _type = "Task" [ 755.263865] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.278558] env[65522]: DEBUG oslo_vmware.api [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113893, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.418486] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375b5b4b-dfeb-4a03-aca5-43c6c3071659 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.428479] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf21fe9-1858-4d9f-85a2-d38878d1852a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.467903] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95208266-a122-4f67-91c1-43294e7ef870 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.480269] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b674ed-96ba-444e-90df-92db21b9ec96 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.496457] env[65522]: DEBUG nova.compute.provider_tree [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.499998] env[65522]: WARNING neutronclient.v2_0.client [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 755.501497] env[65522]: WARNING openstack [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 755.501497] env[65522]: WARNING openstack [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 755.532962] env[65522]: DEBUG oslo_vmware.api [None req-3021d99c-8605-4f06-aaaa-6e41a511fedc tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113891, 'name': PowerOffVM_Task, 'duration_secs': 0.283111} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.533268] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-3021d99c-8605-4f06-aaaa-6e41a511fedc tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 755.533480] env[65522]: DEBUG nova.compute.manager [None req-3021d99c-8605-4f06-aaaa-6e41a511fedc tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 755.534324] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f8a1b6-faf1-4a6b-ad21-ef19c60926e0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.662421] env[65522]: WARNING neutronclient.v2_0.client [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 755.663162] env[65522]: WARNING openstack [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 755.663528] env[65522]: WARNING openstack [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 755.711145] env[65522]: INFO nova.compute.manager [-] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Took 1.66 seconds to deallocate network for instance. [ 755.741640] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113892, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.776289] env[65522]: DEBUG oslo_vmware.api [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113893, 'name': PowerOffVM_Task, 'duration_secs': 0.225871} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.776865] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 755.777216] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 755.777757] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd144af0-e90e-4cce-8777-f3482cf167a9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.891197] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 755.891197] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 755.891197] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Deleting the datastore file [datastore1] 428873b8-1e58-4924-86f3-8feca8f53342 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 755.891197] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d165c763-baa8-474d-b30b-49a08a213489 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.900312] env[65522]: DEBUG oslo_vmware.api [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 755.900312] env[65522]: value = "task-5113895" [ 755.900312] env[65522]: _type = "Task" [ 755.900312] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.914324] env[65522]: DEBUG oslo_vmware.api [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113895, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.000123] env[65522]: DEBUG nova.scheduler.client.report [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 756.047980] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3021d99c-8605-4f06-aaaa-6e41a511fedc tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "7c6f8218-602d-44f3-8012-de5a96972785" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.059s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 756.220287] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.241875] env[65522]: DEBUG oslo_vmware.api [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Task: {'id': task-5113892, 'name': PowerOnVM_Task, 'duration_secs': 0.99914} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.242366] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 756.243083] env[65522]: DEBUG nova.compute.manager [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 756.244649] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c35d4d99-4a98-405f-8026-7be5ad544cd3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.417668] env[65522]: DEBUG oslo_vmware.api [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5113895, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.446109} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.417958] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 756.418221] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 756.418443] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 756.418617] env[65522]: INFO nova.compute.manager [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Took 1.18 seconds to destroy the instance on the hypervisor. [ 756.418945] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 756.419214] env[65522]: DEBUG nova.compute.manager [-] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 756.419662] env[65522]: DEBUG nova.network.neutron [-] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 756.419662] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 756.420490] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.420901] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.458465] env[65522]: DEBUG nova.objects.instance [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'pci_requests' on Instance uuid b2d70982-54e7-459c-a0d8-48bf4b6e4345 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 756.505841] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.758s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 756.505841] env[65522]: DEBUG nova.compute.manager [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 756.509284] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.591s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 756.514039] env[65522]: INFO nova.compute.claims [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.769880] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.963044] env[65522]: DEBUG nova.objects.base [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 756.963300] env[65522]: DEBUG nova.network.neutron [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 756.963636] env[65522]: WARNING neutronclient.v2_0.client [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 756.964122] env[65522]: WARNING neutronclient.v2_0.client [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 756.964659] env[65522]: WARNING openstack [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.965016] env[65522]: WARNING openstack [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.973860] env[65522]: WARNING openstack [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.974353] env[65522]: WARNING openstack [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.017816] env[65522]: DEBUG nova.compute.utils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 757.025026] env[65522]: DEBUG nova.compute.manager [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 757.025026] env[65522]: DEBUG nova.network.neutron [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 757.025280] env[65522]: WARNING neutronclient.v2_0.client [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.025552] env[65522]: WARNING neutronclient.v2_0.client [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.026729] env[65522]: WARNING openstack [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.027489] env[65522]: WARNING openstack [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.140896] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.524966] env[65522]: DEBUG nova.compute.manager [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 757.659842] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f5f2db-8b30-4b5d-b178-ade548b945e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.670084] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88768da5-b89b-406f-8a41-a53fd0c0f3e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.705755] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bfbd5d0-1a1d-45c4-8f4f-a0b6c6c624d7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.715877] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21e03856-b82e-4e62-be1f-5909d43b45ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.733861] env[65522]: DEBUG nova.compute.provider_tree [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.740221] env[65522]: DEBUG nova.policy [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b491939798e4481fb433ffb81f366b25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '068c2387de8c406194d9b1762c7292a7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 758.215784] env[65522]: WARNING neutronclient.v2_0.client [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 758.216508] env[65522]: WARNING openstack [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.216851] env[65522]: WARNING openstack [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.243925] env[65522]: DEBUG nova.scheduler.client.report [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 758.416616] env[65522]: DEBUG nova.policy [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '827a4988bb4e4876b842e6da5674d890', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4517df167bd74537a4c0501b788860e8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 758.440045] env[65522]: DEBUG nova.network.neutron [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Successfully created port: 38d3255d-8d8f-4b7f-927e-5a7330420b08 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 758.536408] env[65522]: DEBUG nova.compute.manager [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 758.547487] env[65522]: DEBUG nova.network.neutron [-] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 758.569669] env[65522]: DEBUG nova.virt.hardware [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 758.569919] env[65522]: DEBUG nova.virt.hardware [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 758.570091] env[65522]: DEBUG nova.virt.hardware [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 758.571123] env[65522]: DEBUG nova.virt.hardware [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 758.571343] env[65522]: DEBUG nova.virt.hardware [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 758.571690] env[65522]: DEBUG nova.virt.hardware [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 758.571932] env[65522]: DEBUG nova.virt.hardware [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 758.572109] env[65522]: DEBUG nova.virt.hardware [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 758.572274] env[65522]: DEBUG nova.virt.hardware [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 758.572433] env[65522]: DEBUG nova.virt.hardware [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 758.572854] env[65522]: DEBUG nova.virt.hardware [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 758.574198] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07432f72-dea4-4f10-a322-279e9345ac0a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.584602] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ba4d41-a4a1-49ca-8284-51fe48ba0e94 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.753195] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.244s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 758.753782] env[65522]: DEBUG nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 758.758687] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.788s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 758.760463] env[65522]: INFO nova.compute.claims [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.052335] env[65522]: INFO nova.compute.manager [-] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Took 2.63 seconds to deallocate network for instance. [ 759.217859] env[65522]: DEBUG nova.network.neutron [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Successfully created port: 03d0a08a-a560-4f8a-8cc9-24211dff931b {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 759.262181] env[65522]: DEBUG nova.compute.utils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 759.263649] env[65522]: DEBUG nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 759.263982] env[65522]: DEBUG nova.network.neutron [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 759.264603] env[65522]: WARNING neutronclient.v2_0.client [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.265020] env[65522]: WARNING neutronclient.v2_0.client [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.266107] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.266986] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.522444] env[65522]: DEBUG nova.network.neutron [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Updating instance_info_cache with network_info: [{"id": "3f4547a2-4a63-4803-a79f-eda70e48dd48", "address": "fa:16:3e:c2:94:99", "network": {"id": "6314277f-6ed5-45e5-950d-2683782849aa", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-246048775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95bca233019f4e2f8bac0dd800392a69", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f4547a2-4a", "ovs_interfaceid": "3f4547a2-4a63-4803-a79f-eda70e48dd48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 759.564050] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 759.777077] env[65522]: DEBUG nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 760.026142] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Releasing lock "refresh_cache-e5d2cc50-4ee6-4e64-9b52-888968a717ca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 760.272235] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e912d72-9259-4139-95f1-79d0b42eb868 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.281217] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d084b2-ba46-4e75-8efa-ad52680893f2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.322386] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e16cd40-e1ad-4b68-942b-29f0a1c6b2df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.333157] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8210c44-884c-420b-912e-27c848d05ba6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.351517] env[65522]: DEBUG nova.compute.provider_tree [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.483422] env[65522]: DEBUG nova.network.neutron [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Successfully updated port: 38d3255d-8d8f-4b7f-927e-5a7330420b08 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 760.540934] env[65522]: DEBUG nova.policy [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a48bd1f8ea6b4c2ab92d4308f986f625', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'acbde84c090e4c0c919728b809c73d02', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 760.822416] env[65522]: DEBUG nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 760.851350] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 760.851603] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 760.851753] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 760.851927] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 760.852082] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 760.852256] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 760.852587] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 760.852765] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 760.852934] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 760.853212] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 760.853310] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 760.854216] env[65522]: DEBUG nova.scheduler.client.report [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 760.858647] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8910085f-33da-4cf0-84bb-3f39f5646145 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.869519] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a772e1f6-ddea-48c5-942c-7c9b9392114e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.987660] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.987785] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 760.988070] env[65522]: DEBUG nova.network.neutron [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 761.035327] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 761.035723] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7fc67f5b-f2a8-402a-8ae7-d4e3690504e1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.045785] env[65522]: DEBUG oslo_vmware.api [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 761.045785] env[65522]: value = "task-5113896" [ 761.045785] env[65522]: _type = "Task" [ 761.045785] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.055553] env[65522]: DEBUG oslo_vmware.api [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113896, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.135282] env[65522]: DEBUG nova.network.neutron [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Successfully updated port: 03d0a08a-a560-4f8a-8cc9-24211dff931b {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 761.188532] env[65522]: DEBUG nova.network.neutron [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Successfully created port: cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 761.363673] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.605s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 761.364243] env[65522]: DEBUG nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 761.366964] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.289s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 761.369286] env[65522]: INFO nova.compute.claims [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.461770] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.462217] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.462440] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.462692] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.462974] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.463223] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.463450] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.463613] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 761.463782] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.492039] env[65522]: WARNING openstack [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 761.492300] env[65522]: WARNING openstack [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 761.558669] env[65522]: DEBUG oslo_vmware.api [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113896, 'name': PowerOnVM_Task, 'duration_secs': 0.46219} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.559078] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 761.559233] env[65522]: DEBUG nova.compute.manager [None req-c4d016f1-aeb2-4f8e-9bc3-382025b05b36 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 761.560601] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90fb27d5-1a93-4690-8133-a05e8e6fadbf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.639131] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Acquiring lock "refresh_cache-ea5b8e29-ebe6-444b-bd2d-029201003ee6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.639394] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Acquired lock "refresh_cache-ea5b8e29-ebe6-444b-bd2d-029201003ee6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 761.639664] env[65522]: DEBUG nova.network.neutron [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 761.874999] env[65522]: DEBUG nova.compute.utils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 761.876432] env[65522]: DEBUG nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 761.876580] env[65522]: DEBUG nova.network.neutron [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 761.876909] env[65522]: WARNING neutronclient.v2_0.client [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 761.877239] env[65522]: WARNING neutronclient.v2_0.client [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 761.878032] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 761.878420] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 761.932264] env[65522]: WARNING nova.network.neutron [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] 1ea0ec37-ea40-4911-b6ad-54375253cc42 already exists in list: networks containing: ['1ea0ec37-ea40-4911-b6ad-54375253cc42']. ignoring it [ 761.966969] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.142711] env[65522]: WARNING openstack [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 762.143108] env[65522]: WARNING openstack [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.159206] env[65522]: DEBUG nova.policy [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a48bd1f8ea6b4c2ab92d4308f986f625', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'acbde84c090e4c0c919728b809c73d02', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 762.390648] env[65522]: DEBUG nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 762.453424] env[65522]: WARNING openstack [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 762.453869] env[65522]: WARNING openstack [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.887284] env[65522]: DEBUG nova.network.neutron [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Successfully updated port: cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 762.920624] env[65522]: DEBUG nova.network.neutron [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 763.010014] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2433b986-6520-484c-b802-261580245768 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.019332] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb91ba5-5ad9-40fa-8c6d-0301023c357e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.054373] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bc1fa45-4166-4e9d-8f31-19d2d7a37121 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.063543] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbba6fe8-1605-43fa-9b27-2ec255e97725 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.079041] env[65522]: DEBUG nova.compute.provider_tree [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.190930] env[65522]: DEBUG nova.network.neutron [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Successfully created port: 3d5e8380-9410-49a2-873c-962835dabfaa {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 763.274177] env[65522]: WARNING openstack [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.275032] env[65522]: WARNING openstack [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 763.390360] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "refresh_cache-4238d766-2a5d-40d7-b2f6-5cb9a5354760" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.390551] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired lock "refresh_cache-4238d766-2a5d-40d7-b2f6-5cb9a5354760" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.390777] env[65522]: DEBUG nova.network.neutron [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 763.406710] env[65522]: DEBUG nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 763.439729] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 763.439863] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 763.440044] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 763.440246] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 763.440391] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 763.440676] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 763.440775] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 763.440904] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 763.441090] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 763.441246] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 763.441431] env[65522]: DEBUG nova.virt.hardware [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 763.442753] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a96f54a4-cd8f-4f83-83d7-a92b706413b8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.452932] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d8b315-d8f3-4c49-8920-d7bf9c3523d1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.582234] env[65522]: DEBUG nova.scheduler.client.report [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 763.894656] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 763.895074] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.088772] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.722s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 764.089414] env[65522]: DEBUG nova.compute.manager [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 764.092696] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.194s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 764.094660] env[65522]: INFO nova.compute.claims [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 764.309875] env[65522]: WARNING neutronclient.v2_0.client [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 764.310673] env[65522]: WARNING openstack [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.311245] env[65522]: WARNING openstack [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.602900] env[65522]: DEBUG nova.compute.utils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 764.605001] env[65522]: DEBUG nova.compute.manager [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 764.605212] env[65522]: DEBUG nova.network.neutron [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 764.605571] env[65522]: WARNING neutronclient.v2_0.client [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 764.605846] env[65522]: WARNING neutronclient.v2_0.client [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 764.606553] env[65522]: WARNING openstack [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.606886] env[65522]: WARNING openstack [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.844301] env[65522]: DEBUG nova.network.neutron [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Successfully updated port: 3d5e8380-9410-49a2-873c-962835dabfaa {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 764.906405] env[65522]: DEBUG nova.network.neutron [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 764.989309] env[65522]: WARNING neutronclient.v2_0.client [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 764.989754] env[65522]: WARNING openstack [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.990160] env[65522]: WARNING openstack [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.067219] env[65522]: DEBUG nova.policy [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2109e40451943a3884012e51bf118f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b5c14a856d454048856539895f220c2c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 765.105737] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.106138] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.116631] env[65522]: DEBUG nova.compute.manager [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 765.271041] env[65522]: WARNING openstack [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.271284] env[65522]: WARNING openstack [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.350962] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "refresh_cache-1e8088ef-6354-40a0-95ee-06732e7f088f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.350962] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired lock "refresh_cache-1e8088ef-6354-40a0-95ee-06732e7f088f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 765.350962] env[65522]: DEBUG nova.network.neutron [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 765.564477] env[65522]: DEBUG nova.network.neutron [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Successfully created port: 652babe2-668c-4364-92c2-8ceaba23c0ac {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 765.613226] env[65522]: DEBUG nova.network.neutron [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Updating instance_info_cache with network_info: [{"id": "03d0a08a-a560-4f8a-8cc9-24211dff931b", "address": "fa:16:3e:d8:09:37", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03d0a08a-a5", "ovs_interfaceid": "03d0a08a-a560-4f8a-8cc9-24211dff931b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 765.630690] env[65522]: WARNING neutronclient.v2_0.client [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 765.632435] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.632435] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.732142] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457f95ca-9831-45ac-b568-dbf23d4496fb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.744759] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ac5ed8-87b7-4ad1-ac45-75c5b1f3ca26 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.788203] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4bb2a40-27dc-481e-a869-faea47186292 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.798204] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d85c17c-5d14-4fcc-b4df-02ddb519668c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.814517] env[65522]: DEBUG nova.compute.provider_tree [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.860636] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.860636] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.126056] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Releasing lock "refresh_cache-ea5b8e29-ebe6-444b-bd2d-029201003ee6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 766.127262] env[65522]: DEBUG nova.compute.manager [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Instance network_info: |[{"id": "03d0a08a-a560-4f8a-8cc9-24211dff931b", "address": "fa:16:3e:d8:09:37", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03d0a08a-a5", "ovs_interfaceid": "03d0a08a-a560-4f8a-8cc9-24211dff931b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 766.127577] env[65522]: DEBUG nova.compute.manager [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 766.129755] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:09:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03d0a08a-a560-4f8a-8cc9-24211dff931b', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 766.137161] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Creating folder: Project (4517df167bd74537a4c0501b788860e8). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 766.137857] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f2980e97-c116-4906-9b37-97ba7f6ef1c3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.149933] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Created folder: Project (4517df167bd74537a4c0501b788860e8) in parent group-v994660. [ 766.150143] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Creating folder: Instances. Parent ref: group-v994765. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 766.150404] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-694ca2ca-f48a-4216-a810-5df0fc75dc61 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.162160] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Created folder: Instances in parent group-v994765. [ 766.162418] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 766.162624] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 766.162846] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6723a81d-5590-42ea-8a05-916a9092d874 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.185085] env[65522]: DEBUG nova.virt.hardware [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 766.185537] env[65522]: DEBUG nova.virt.hardware [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 766.185926] env[65522]: DEBUG nova.virt.hardware [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 766.186178] env[65522]: DEBUG nova.virt.hardware [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 766.186390] env[65522]: DEBUG nova.virt.hardware [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 766.186577] env[65522]: DEBUG nova.virt.hardware [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 766.186962] env[65522]: DEBUG nova.virt.hardware [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.187222] env[65522]: DEBUG nova.virt.hardware [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 766.187439] env[65522]: DEBUG nova.virt.hardware [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 766.187632] env[65522]: DEBUG nova.virt.hardware [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 766.187983] env[65522]: DEBUG nova.virt.hardware [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 766.189079] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e37f772-c0c6-4e75-a408-eb408d2a236b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.200087] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4438f6-0827-4a03-99ef-a75020581b57 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.207333] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 766.207333] env[65522]: value = "task-5113899" [ 766.207333] env[65522]: _type = "Task" [ 766.207333] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.226881] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113899, 'name': CreateVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.318612] env[65522]: DEBUG nova.scheduler.client.report [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 766.682091] env[65522]: WARNING neutronclient.v2_0.client [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 766.682841] env[65522]: WARNING openstack [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.683260] env[65522]: WARNING openstack [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.715774] env[65522]: DEBUG nova.network.neutron [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 766.725248] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113899, 'name': CreateVM_Task, 'duration_secs': 0.367238} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.725392] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 766.725900] env[65522]: WARNING neutronclient.v2_0.client [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 766.726304] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.726416] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 766.726927] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 766.727226] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1be7d6f6-501b-4109-b354-78667e24e931 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.733759] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Waiting for the task: (returnval){ [ 766.733759] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527632d9-b4c6-719c-2391-efa970457c99" [ 766.733759] env[65522]: _type = "Task" [ 766.733759] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.744539] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527632d9-b4c6-719c-2391-efa970457c99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.756944] env[65522]: DEBUG nova.network.neutron [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Updating instance_info_cache with network_info: [{"id": "cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c", "address": "fa:16:3e:2c:05:d9", "network": {"id": "0ab23355-59a7-409b-89f3-089fbd03db5e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-960579569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "acbde84c090e4c0c919728b809c73d02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf5b96ab-5c", "ovs_interfaceid": "cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 766.824490] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.732s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 766.825034] env[65522]: DEBUG nova.compute.manager [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 766.828511] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.166s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 766.833021] env[65522]: INFO nova.compute.claims [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.840998] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.841396] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.880016] env[65522]: DEBUG nova.network.neutron [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Updating instance_info_cache with network_info: [{"id": "d24163e8-4f08-4908-9307-c4edb0d4ffb7", "address": "fa:16:3e:e2:90:31", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd24163e8-4f", "ovs_interfaceid": "d24163e8-4f08-4908-9307-c4edb0d4ffb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "38d3255d-8d8f-4b7f-927e-5a7330420b08", "address": "fa:16:3e:76:d8:87", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38d3255d-8d", "ovs_interfaceid": "38d3255d-8d8f-4b7f-927e-5a7330420b08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 767.001980] env[65522]: WARNING neutronclient.v2_0.client [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 767.002641] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 767.002893] env[65522]: WARNING openstack [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 767.124994] env[65522]: DEBUG nova.network.neutron [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Updating instance_info_cache with network_info: [{"id": "3d5e8380-9410-49a2-873c-962835dabfaa", "address": "fa:16:3e:1f:ae:89", "network": {"id": "0ab23355-59a7-409b-89f3-089fbd03db5e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-960579569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acbde84c090e4c0c919728b809c73d02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d5e8380-94", "ovs_interfaceid": "3d5e8380-9410-49a2-873c-962835dabfaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 767.170916] env[65522]: DEBUG nova.compute.manager [req-9d3df41d-5626-4d1c-a2bf-2402d61e1a07 req-72095701-968a-4ccf-9cb9-45893c45b0a7 service nova] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Received event network-vif-deleted-a002c980-b974-432b-aa4e-e8613c617e8e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 767.247627] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527632d9-b4c6-719c-2391-efa970457c99, 'name': SearchDatastore_Task, 'duration_secs': 0.012253} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.248112] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 767.248474] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 767.248849] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.249065] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 767.249264] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 767.249554] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c58bfbe2-91b9-45aa-88f5-4bb6ccece495 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.259612] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Releasing lock "refresh_cache-4238d766-2a5d-40d7-b2f6-5cb9a5354760" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 767.259995] env[65522]: DEBUG nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Instance network_info: |[{"id": "cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c", "address": "fa:16:3e:2c:05:d9", "network": {"id": "0ab23355-59a7-409b-89f3-089fbd03db5e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-960579569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "acbde84c090e4c0c919728b809c73d02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf5b96ab-5c", "ovs_interfaceid": "cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 767.260528] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:05:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4712af2-45ef-4652-8d2c-482ec70056d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 767.269591] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Creating folder: Project (acbde84c090e4c0c919728b809c73d02). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 767.269996] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7e67310f-427b-4ed7-8529-e989bdfd8546 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.273172] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 767.273342] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 767.274109] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eeeaf6d3-f11b-44f8-813c-0c681c28af89 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.282067] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Waiting for the task: (returnval){ [ 767.282067] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520ff00e-ae62-da13-06fa-4f56bb34175f" [ 767.282067] env[65522]: _type = "Task" [ 767.282067] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.288293] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Created folder: Project (acbde84c090e4c0c919728b809c73d02) in parent group-v994660. [ 767.288559] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Creating folder: Instances. Parent ref: group-v994768. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 767.289208] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b1024ec-3a29-4fe5-b078-6260a197d327 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.300396] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520ff00e-ae62-da13-06fa-4f56bb34175f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.310885] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Created folder: Instances in parent group-v994768. [ 767.311191] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 767.311410] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 767.311640] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-343e5e63-aa0e-48a1-abc8-94ea9f78d0fa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.335411] env[65522]: DEBUG nova.compute.utils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 767.338760] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 767.338760] env[65522]: value = "task-5113902" [ 767.338760] env[65522]: _type = "Task" [ 767.338760] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.343174] env[65522]: DEBUG nova.compute.manager [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 767.343174] env[65522]: DEBUG nova.network.neutron [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 767.343174] env[65522]: WARNING neutronclient.v2_0.client [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 767.343174] env[65522]: WARNING neutronclient.v2_0.client [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 767.343174] env[65522]: WARNING openstack [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 767.343174] env[65522]: WARNING openstack [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 767.361670] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113902, 'name': CreateVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.386223] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 767.387973] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.387973] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 767.389307] env[65522]: DEBUG nova.network.neutron [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Successfully updated port: 652babe2-668c-4364-92c2-8ceaba23c0ac {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 767.391263] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b371f3d9-76e1-4472-ab54-0f220aff535a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.398321] env[65522]: DEBUG nova.policy [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2109e40451943a3884012e51bf118f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b5c14a856d454048856539895f220c2c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 767.416427] env[65522]: DEBUG nova.virt.hardware [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 767.416427] env[65522]: DEBUG nova.virt.hardware [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 767.416427] env[65522]: DEBUG nova.virt.hardware [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 767.416699] env[65522]: DEBUG nova.virt.hardware [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 767.416699] env[65522]: DEBUG nova.virt.hardware [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 767.416805] env[65522]: DEBUG nova.virt.hardware [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 767.417014] env[65522]: DEBUG nova.virt.hardware [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.417168] env[65522]: DEBUG nova.virt.hardware [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 767.417326] env[65522]: DEBUG nova.virt.hardware [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 767.417486] env[65522]: DEBUG nova.virt.hardware [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 767.417983] env[65522]: DEBUG nova.virt.hardware [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 767.424657] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Reconfiguring VM to attach interface {{(pid=65522) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 767.425808] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4ecea9e-eeb6-4594-821e-7f08a8ac1461 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.450873] env[65522]: DEBUG oslo_vmware.api [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 767.450873] env[65522]: value = "task-5113903" [ 767.450873] env[65522]: _type = "Task" [ 767.450873] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.461813] env[65522]: DEBUG oslo_vmware.api [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113903, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.628322] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Releasing lock "refresh_cache-1e8088ef-6354-40a0-95ee-06732e7f088f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 767.628741] env[65522]: DEBUG nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Instance network_info: |[{"id": "3d5e8380-9410-49a2-873c-962835dabfaa", "address": "fa:16:3e:1f:ae:89", "network": {"id": "0ab23355-59a7-409b-89f3-089fbd03db5e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-960579569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acbde84c090e4c0c919728b809c73d02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d5e8380-94", "ovs_interfaceid": "3d5e8380-9410-49a2-873c-962835dabfaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 767.629264] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:ae:89', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4712af2-45ef-4652-8d2c-482ec70056d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d5e8380-9410-49a2-873c-962835dabfaa', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 767.638138] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 767.638265] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 767.638522] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-caca558f-84b7-4faf-b179-27208f6aa1aa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.660894] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 767.660894] env[65522]: value = "task-5113904" [ 767.660894] env[65522]: _type = "Task" [ 767.660894] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.670402] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113904, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.707731] env[65522]: DEBUG nova.network.neutron [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Successfully created port: f33077f4-9852-477a-ae48-9e86ad80fdce {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 767.796517] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520ff00e-ae62-da13-06fa-4f56bb34175f, 'name': SearchDatastore_Task, 'duration_secs': 0.034773} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.797717] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69a04ed0-cfae-4222-8673-350bbe461812 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.806651] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Waiting for the task: (returnval){ [ 767.806651] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521b43e0-c9e2-5402-e68e-0beb569ef971" [ 767.806651] env[65522]: _type = "Task" [ 767.806651] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.817734] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521b43e0-c9e2-5402-e68e-0beb569ef971, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.853534] env[65522]: DEBUG nova.compute.manager [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 767.871918] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113902, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.901594] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "refresh_cache-bb20d99f-fc49-47bf-a6b7-74ea966b71eb" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.901848] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "refresh_cache-bb20d99f-fc49-47bf-a6b7-74ea966b71eb" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 767.901898] env[65522]: DEBUG nova.network.neutron [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 767.965472] env[65522]: DEBUG oslo_vmware.api [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.174454] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113904, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.317842] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521b43e0-c9e2-5402-e68e-0beb569ef971, 'name': SearchDatastore_Task, 'duration_secs': 0.038922} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.320729] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 768.321041] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] ea5b8e29-ebe6-444b-bd2d-029201003ee6/ea5b8e29-ebe6-444b-bd2d-029201003ee6.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 768.321589] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d59a8ab8-fd9a-48dc-bf5a-b7658f40f589 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.329340] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Waiting for the task: (returnval){ [ 768.329340] env[65522]: value = "task-5113905" [ 768.329340] env[65522]: _type = "Task" [ 768.329340] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.342630] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113905, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.361714] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113902, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.406081] env[65522]: WARNING openstack [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.406493] env[65522]: WARNING openstack [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.414301] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2136f485-12c8-44f5-b34f-32d06588c5d0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.423193] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f7706e-b7ed-4a0a-b67e-d56ef0a23ddb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.455931] env[65522]: DEBUG nova.network.neutron [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 768.461381] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc58325-6818-43b9-8f1d-3a9d41a20036 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.474133] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95919087-ae45-4269-b05c-b02f4319e118 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.478369] env[65522]: DEBUG oslo_vmware.api [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.490198] env[65522]: DEBUG nova.compute.provider_tree [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.495647] env[65522]: WARNING openstack [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.495647] env[65522]: WARNING openstack [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.549379] env[65522]: WARNING neutronclient.v2_0.client [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.550309] env[65522]: WARNING openstack [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.550703] env[65522]: WARNING openstack [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.677353] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113904, 'name': CreateVM_Task, 'duration_secs': 0.726559} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.677704] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 768.678719] env[65522]: WARNING neutronclient.v2_0.client [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.679349] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.679655] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.680111] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 768.680491] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ba518f2-399e-4f9a-beb4-cd7926da160d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.690543] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 768.690543] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cbcc7e-b791-28cd-21a1-51ce877a480f" [ 768.690543] env[65522]: _type = "Task" [ 768.690543] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.706267] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cbcc7e-b791-28cd-21a1-51ce877a480f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.724023] env[65522]: DEBUG nova.network.neutron [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Updating instance_info_cache with network_info: [{"id": "652babe2-668c-4364-92c2-8ceaba23c0ac", "address": "fa:16:3e:fe:e2:bf", "network": {"id": "dc9d8462-3539-4cc5-b21a-e1136a8df8e6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1580876213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "b5c14a856d454048856539895f220c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap652babe2-66", "ovs_interfaceid": "652babe2-668c-4364-92c2-8ceaba23c0ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 768.778296] env[65522]: DEBUG nova.compute.manager [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 768.779063] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db8d433-7ae2-48d3-81fc-b69e87b9968a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.843831] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113905, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.866153] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113902, 'name': CreateVM_Task, 'duration_secs': 1.424763} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.866374] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 768.867060] env[65522]: WARNING neutronclient.v2_0.client [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.867658] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.870533] env[65522]: DEBUG nova.compute.manager [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 768.907845] env[65522]: DEBUG nova.virt.hardware [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 768.907845] env[65522]: DEBUG nova.virt.hardware [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 768.907845] env[65522]: DEBUG nova.virt.hardware [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 768.907845] env[65522]: DEBUG nova.virt.hardware [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 768.907845] env[65522]: DEBUG nova.virt.hardware [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 768.907845] env[65522]: DEBUG nova.virt.hardware [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 768.907845] env[65522]: DEBUG nova.virt.hardware [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 768.908171] env[65522]: DEBUG nova.virt.hardware [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 768.908171] env[65522]: DEBUG nova.virt.hardware [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 768.908271] env[65522]: DEBUG nova.virt.hardware [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 768.908431] env[65522]: DEBUG nova.virt.hardware [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 768.909885] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60941705-fa2e-445b-8b92-1b4933188cad {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.921764] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4636ff3-f97e-4461-a0e5-c468a0d540d4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.969503] env[65522]: DEBUG oslo_vmware.api [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.002427] env[65522]: DEBUG nova.scheduler.client.report [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 769.134231] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Acquiring lock "d2fdffb8-76d1-4b74-819f-25eb29ca91ce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 769.134487] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Lock "d2fdffb8-76d1-4b74-819f-25eb29ca91ce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 769.203209] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cbcc7e-b791-28cd-21a1-51ce877a480f, 'name': SearchDatastore_Task, 'duration_secs': 0.062794} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.203551] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 769.203833] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 769.204138] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.204503] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 769.204503] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 769.204754] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 769.205500] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 769.205500] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4fbec1b-4dd3-4c42-9083-c46273560967 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.207505] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d323b00-70b5-4088-b606-8c6e8c35d89e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.214096] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 769.214096] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5286763c-9e7c-6919-6cb5-5f2459c5504e" [ 769.214096] env[65522]: _type = "Task" [ 769.214096] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.219213] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 769.219367] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 769.220570] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eca5c21e-ff94-4677-b457-8197f7c41ef4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.226921] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "refresh_cache-bb20d99f-fc49-47bf-a6b7-74ea966b71eb" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 769.227135] env[65522]: DEBUG nova.compute.manager [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Instance network_info: |[{"id": "652babe2-668c-4364-92c2-8ceaba23c0ac", "address": "fa:16:3e:fe:e2:bf", "network": {"id": "dc9d8462-3539-4cc5-b21a-e1136a8df8e6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1580876213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "b5c14a856d454048856539895f220c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap652babe2-66", "ovs_interfaceid": "652babe2-668c-4364-92c2-8ceaba23c0ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 769.227465] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5286763c-9e7c-6919-6cb5-5f2459c5504e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.228329] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:e2:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '652babe2-668c-4364-92c2-8ceaba23c0ac', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 769.235998] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Creating folder: Project (b5c14a856d454048856539895f220c2c). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 769.237745] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c2a0feb8-cc05-469a-8d6a-857562fb916c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.239764] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 769.239764] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fc2468-99f0-ab66-18cb-75f56add3530" [ 769.239764] env[65522]: _type = "Task" [ 769.239764] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.251177] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fc2468-99f0-ab66-18cb-75f56add3530, 'name': SearchDatastore_Task, 'duration_secs': 0.011986} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.253691] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Created folder: Project (b5c14a856d454048856539895f220c2c) in parent group-v994660. [ 769.254054] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Creating folder: Instances. Parent ref: group-v994772. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 769.254712] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26ce0cd2-a379-4d71-a5de-71aa0244d196 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.258442] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-76b916ca-7925-43a4-b196-fc2505a4d60f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.264124] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 769.264124] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528629c7-2423-2ba7-1494-a1e3635d5aff" [ 769.264124] env[65522]: _type = "Task" [ 769.264124] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.271731] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Created folder: Instances in parent group-v994772. [ 769.271731] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 769.272286] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 769.272615] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89e09051-75f7-4e0e-91c9-9a3fd2a19ce4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.293589] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528629c7-2423-2ba7-1494-a1e3635d5aff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.295320] env[65522]: INFO nova.compute.manager [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] instance snapshotting [ 769.298669] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17a3c39a-98e5-499e-a265-dba76c943ed9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.304348] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 769.304348] env[65522]: value = "task-5113908" [ 769.304348] env[65522]: _type = "Task" [ 769.304348] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.324851] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8464c7-0dbc-4af0-9c35-8f1ffafa7531 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.332122] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113908, 'name': CreateVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.344700] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113905, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.667993} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.345393] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] ea5b8e29-ebe6-444b-bd2d-029201003ee6/ea5b8e29-ebe6-444b-bd2d-029201003ee6.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 769.345603] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 769.345770] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b3e5873a-4b10-4cc1-8064-aa3e60bf107d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.356916] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Waiting for the task: (returnval){ [ 769.356916] env[65522]: value = "task-5113909" [ 769.356916] env[65522]: _type = "Task" [ 769.356916] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.369662] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113909, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.423163] env[65522]: DEBUG nova.network.neutron [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Successfully updated port: f33077f4-9852-477a-ae48-9e86ad80fdce {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 769.469195] env[65522]: DEBUG oslo_vmware.api [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113903, 'name': ReconfigVM_Task, 'duration_secs': 1.786197} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.469665] env[65522]: WARNING neutronclient.v2_0.client [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 769.469912] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 769.470148] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Reconfigured VM to attach interface {{(pid=65522) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 769.507859] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.679s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 769.508450] env[65522]: DEBUG nova.compute.manager [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 769.511626] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.521s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 769.511874] env[65522]: DEBUG nova.objects.instance [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Lazy-loading 'resources' on Instance uuid 6abf58ca-ad6b-4fe8-844a-ca33e62a844d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 769.637678] env[65522]: DEBUG nova.compute.manager [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 769.725428] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5286763c-9e7c-6919-6cb5-5f2459c5504e, 'name': SearchDatastore_Task, 'duration_secs': 0.012272} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.725800] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 769.725895] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 769.726151] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.777438] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528629c7-2423-2ba7-1494-a1e3635d5aff, 'name': SearchDatastore_Task, 'duration_secs': 0.019614} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.778618] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 769.778618] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 1e8088ef-6354-40a0-95ee-06732e7f088f/1e8088ef-6354-40a0-95ee-06732e7f088f.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 769.778618] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 769.778953] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 769.779298] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c0c5688-98e1-401b-9ddc-8763598280a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.782156] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f771d5a6-e08a-44d9-979c-25987dfe8e6d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.791411] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 769.791411] env[65522]: value = "task-5113910" [ 769.791411] env[65522]: _type = "Task" [ 769.791411] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.796066] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 769.796150] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 769.797399] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12519cad-60fc-4a1a-8e17-f769ece86fcc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.803424] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113910, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.808477] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 769.808477] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5236bf0d-6a72-4968-b8f3-29b9efa5a705" [ 769.808477] env[65522]: _type = "Task" [ 769.808477] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.816071] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113908, 'name': CreateVM_Task, 'duration_secs': 0.403207} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.816347] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 769.816772] env[65522]: WARNING neutronclient.v2_0.client [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 769.817275] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.817460] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 769.817784] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 769.818119] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aec434fc-36a9-4f3b-ac52-81304664ea4a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.823384] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5236bf0d-6a72-4968-b8f3-29b9efa5a705, 'name': SearchDatastore_Task, 'duration_secs': 0.010769} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.824693] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8cfab9c-90db-4338-8c19-61e9d70d5cbc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.828847] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 769.828847] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a0870f-9237-494c-affa-38698f39d4d4" [ 769.828847] env[65522]: _type = "Task" [ 769.828847] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.834760] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 769.834760] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a887cb-ea9d-5e71-4f8f-d168aee7fd20" [ 769.834760] env[65522]: _type = "Task" [ 769.834760] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.842091] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a0870f-9237-494c-affa-38698f39d4d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.849521] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 769.850297] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a887cb-ea9d-5e71-4f8f-d168aee7fd20, 'name': SearchDatastore_Task, 'duration_secs': 0.012011} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.850297] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-91bfc9bd-c0d0-4c6e-8364-9e512cc11990 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.852305] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 769.852502] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 4238d766-2a5d-40d7-b2f6-5cb9a5354760/4238d766-2a5d-40d7-b2f6-5cb9a5354760.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 769.853296] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6937f8a6-fb3e-46e4-90eb-7b7a2aef9c84 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.864146] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 769.864146] env[65522]: value = "task-5113911" [ 769.864146] env[65522]: _type = "Task" [ 769.864146] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.866766] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 769.866766] env[65522]: value = "task-5113912" [ 769.866766] env[65522]: _type = "Task" [ 769.866766] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.877804] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113909, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.12158} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.878350] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 769.879708] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0565ac-a366-4b10-9a93-1d04d3791fd5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.892400] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113911, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.892845] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113912, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.915083] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] ea5b8e29-ebe6-444b-bd2d-029201003ee6/ea5b8e29-ebe6-444b-bd2d-029201003ee6.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 769.915471] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bcb0a2ec-c631-4efb-96bc-b1d2abf5f4aa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.932903] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "refresh_cache-24f520ec-6a56-4f17-9ae4-2c856bf99582" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.933192] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "refresh_cache-24f520ec-6a56-4f17-9ae4-2c856bf99582" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 769.933296] env[65522]: DEBUG nova.network.neutron [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 769.936197] env[65522]: DEBUG nova.compute.manager [req-111a7a3d-4c6a-430c-8568-dc505ebac951 req-08e2d55e-2243-46d5-b82c-3f4927117475 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Received event network-vif-plugged-38d3255d-8d8f-4b7f-927e-5a7330420b08 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 769.936521] env[65522]: DEBUG oslo_concurrency.lockutils [req-111a7a3d-4c6a-430c-8568-dc505ebac951 req-08e2d55e-2243-46d5-b82c-3f4927117475 service nova] Acquiring lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 769.938945] env[65522]: DEBUG oslo_concurrency.lockutils [req-111a7a3d-4c6a-430c-8568-dc505ebac951 req-08e2d55e-2243-46d5-b82c-3f4927117475 service nova] Lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 769.939259] env[65522]: DEBUG oslo_concurrency.lockutils [req-111a7a3d-4c6a-430c-8568-dc505ebac951 req-08e2d55e-2243-46d5-b82c-3f4927117475 service nova] Lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 769.939550] env[65522]: DEBUG nova.compute.manager [req-111a7a3d-4c6a-430c-8568-dc505ebac951 req-08e2d55e-2243-46d5-b82c-3f4927117475 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] No waiting events found dispatching network-vif-plugged-38d3255d-8d8f-4b7f-927e-5a7330420b08 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 769.939880] env[65522]: WARNING nova.compute.manager [req-111a7a3d-4c6a-430c-8568-dc505ebac951 req-08e2d55e-2243-46d5-b82c-3f4927117475 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Received unexpected event network-vif-plugged-38d3255d-8d8f-4b7f-927e-5a7330420b08 for instance with vm_state active and task_state None. [ 769.948126] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Waiting for the task: (returnval){ [ 769.948126] env[65522]: value = "task-5113913" [ 769.948126] env[65522]: _type = "Task" [ 769.948126] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.961369] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113913, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.975283] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f908f0a-34d6-45f3-862a-e646789026ae tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-b2d70982-54e7-459c-a0d8-48bf4b6e4345-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 14.818s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.020186] env[65522]: DEBUG nova.compute.utils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 770.028105] env[65522]: DEBUG nova.compute.manager [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 770.028193] env[65522]: DEBUG nova.network.neutron [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 770.028664] env[65522]: WARNING neutronclient.v2_0.client [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.029671] env[65522]: WARNING neutronclient.v2_0.client [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.030251] env[65522]: WARNING openstack [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.030771] env[65522]: WARNING openstack [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.114840] env[65522]: DEBUG nova.policy [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e74c132f591d4c6fbe65a75c1e1e4a24', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ecccb656b0d4c96b40b200cdcddbad5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 770.157000] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.157336] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.180858] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.307104] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113910, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.348668] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a0870f-9237-494c-affa-38698f39d4d4, 'name': SearchDatastore_Task, 'duration_secs': 0.013019} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.349080] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 770.349392] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 770.349638] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.349781] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 770.349972] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 770.350284] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1fc17371-5899-4e91-9855-2956fc5d92ba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.370801] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 770.371034] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 770.380019] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59f07058-34a4-4940-b173-5890176707fa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.395358] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113912, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.403881] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113911, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.403881] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 770.403881] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522f7197-20b4-fda6-b81c-14d90e7a7391" [ 770.403881] env[65522]: _type = "Task" [ 770.403881] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.421206] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522f7197-20b4-fda6-b81c-14d90e7a7391, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.447774] env[65522]: WARNING openstack [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.448184] env[65522]: WARNING openstack [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.467715] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113913, 'name': ReconfigVM_Task, 'duration_secs': 0.485207} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.468023] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Reconfigured VM instance instance-00000026 to attach disk [datastore2] ea5b8e29-ebe6-444b-bd2d-029201003ee6/ea5b8e29-ebe6-444b-bd2d-029201003ee6.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 770.471821] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef050ffa-1ef2-4394-9c9b-64cb52b20cd5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.483291] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Waiting for the task: (returnval){ [ 770.483291] env[65522]: value = "task-5113914" [ 770.483291] env[65522]: _type = "Task" [ 770.483291] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.502644] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113914, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.529365] env[65522]: DEBUG nova.network.neutron [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Successfully created port: 9446c1bb-b443-4a52-879d-5fa6cb83f0dc {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 770.532736] env[65522]: DEBUG nova.compute.manager [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 770.588020] env[65522]: DEBUG nova.network.neutron [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 770.620046] env[65522]: WARNING openstack [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.620604] env[65522]: WARNING openstack [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.669166] env[65522]: DEBUG nova.compute.manager [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 770.673553] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.673553] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.673901] env[65522]: DEBUG nova.compute.manager [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Going to confirm migration 1 {{(pid=65522) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 770.709810] env[65522]: WARNING neutronclient.v2_0.client [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.709810] env[65522]: WARNING openstack [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.709810] env[65522]: WARNING openstack [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.807341] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113910, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.606616} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.807872] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 1e8088ef-6354-40a0-95ee-06732e7f088f/1e8088ef-6354-40a0-95ee-06732e7f088f.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 770.807969] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 770.808293] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de82e40e-c603-469b-bbca-ace2a00c20b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.822854] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 770.822854] env[65522]: value = "task-5113915" [ 770.822854] env[65522]: _type = "Task" [ 770.822854] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.833230] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113915, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.835301] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea684902-c19d-49c1-a028-16ffb7a4ab8e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.845811] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb4c2fd0-8194-41d8-bf06-6c5ff64f6520 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.888154] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9f2bed-4936-4338-936e-b6223e2c2cd4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.897060] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113911, 'name': CreateSnapshot_Task, 'duration_secs': 0.705282} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.903001] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 770.904238] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113912, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.959023} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.904238] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-402488b9-1927-49b9-bdf2-78547dee711c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.907496] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 4238d766-2a5d-40d7-b2f6-5cb9a5354760/4238d766-2a5d-40d7-b2f6-5cb9a5354760.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 770.907686] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 770.909124] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5943ecb-4644-449f-83fb-bc96b66809ae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.913274] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e57e63fc-b7cb-41c5-a3de-9c95f469c41f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.936687] env[65522]: DEBUG nova.compute.provider_tree [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.942482] env[65522]: DEBUG nova.network.neutron [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Updating instance_info_cache with network_info: [{"id": "f33077f4-9852-477a-ae48-9e86ad80fdce", "address": "fa:16:3e:43:06:9c", "network": {"id": "dc9d8462-3539-4cc5-b21a-e1136a8df8e6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1580876213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "b5c14a856d454048856539895f220c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf33077f4-98", "ovs_interfaceid": "f33077f4-9852-477a-ae48-9e86ad80fdce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 770.948081] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 770.948081] env[65522]: value = "task-5113916" [ 770.948081] env[65522]: _type = "Task" [ 770.948081] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.948365] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522f7197-20b4-fda6-b81c-14d90e7a7391, 'name': SearchDatastore_Task, 'duration_secs': 0.025754} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.955022] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a857dc0-2ebe-4212-8539-072b651c47bb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.967089] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 770.967089] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ac539a-8332-9705-6d95-20ea80924b3a" [ 770.967089] env[65522]: _type = "Task" [ 770.967089] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.967362] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113916, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.981012] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ac539a-8332-9705-6d95-20ea80924b3a, 'name': SearchDatastore_Task, 'duration_secs': 0.011885} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.981351] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 770.981572] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] bb20d99f-fc49-47bf-a6b7-74ea966b71eb/bb20d99f-fc49-47bf-a6b7-74ea966b71eb.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 770.981854] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b65a1cb-5e21-4369-b1e8-cce4b32f78e2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.995439] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113914, 'name': Rename_Task, 'duration_secs': 0.195728} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.997435] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 770.997826] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 770.997826] env[65522]: value = "task-5113917" [ 770.997826] env[65522]: _type = "Task" [ 770.997826] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.998157] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed75e59c-9515-4f6b-b6a9-3ce6be4b5c05 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.010963] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113917, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.012651] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Waiting for the task: (returnval){ [ 771.012651] env[65522]: value = "task-5113918" [ 771.012651] env[65522]: _type = "Task" [ 771.012651] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.022806] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113918, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.180087] env[65522]: WARNING neutronclient.v2_0.client [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.203762] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 771.228222] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquiring lock "6e10e846-cdb7-458e-b511-18ab1742228e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 771.228737] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Lock "6e10e846-cdb7-458e-b511-18ab1742228e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 771.229133] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquiring lock "6e10e846-cdb7-458e-b511-18ab1742228e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 771.229366] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Lock "6e10e846-cdb7-458e-b511-18ab1742228e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 771.229543] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Lock "6e10e846-cdb7-458e-b511-18ab1742228e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 771.233889] env[65522]: INFO nova.compute.manager [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Terminating instance [ 771.334546] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113915, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092015} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.334894] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 771.335928] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-180b7a5a-e93b-4f94-ac44-1cf9aa7e6d24 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.361164] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 1e8088ef-6354-40a0-95ee-06732e7f088f/1e8088ef-6354-40a0-95ee-06732e7f088f.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 771.361618] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-15167d8c-1a90-4324-ada2-975a910c9bb0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.383692] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 771.383692] env[65522]: value = "task-5113919" [ 771.383692] env[65522]: _type = "Task" [ 771.383692] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.397339] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113919, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.447896] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 771.448327] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a2657bf1-b5d6-45bc-bac1-76e996cf84e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.452756] env[65522]: DEBUG nova.scheduler.client.report [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 771.456600] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "refresh_cache-24f520ec-6a56-4f17-9ae4-2c856bf99582" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 771.456776] env[65522]: DEBUG nova.compute.manager [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Instance network_info: |[{"id": "f33077f4-9852-477a-ae48-9e86ad80fdce", "address": "fa:16:3e:43:06:9c", "network": {"id": "dc9d8462-3539-4cc5-b21a-e1136a8df8e6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1580876213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "b5c14a856d454048856539895f220c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf33077f4-98", "ovs_interfaceid": "f33077f4-9852-477a-ae48-9e86ad80fdce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 771.457596] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:06:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f33077f4-9852-477a-ae48-9e86ad80fdce', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 771.465651] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 771.470571] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 771.471908] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-10ed186f-6e66-40a1-9d8e-9ac52af12a6c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.492820] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 771.492820] env[65522]: value = "task-5113920" [ 771.492820] env[65522]: _type = "Task" [ 771.492820] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.498571] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113916, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074898} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.502794] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 771.505509] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9337609f-114c-46f9-8257-bba66fa8a2f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.508861] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 771.508861] env[65522]: value = "task-5113921" [ 771.508861] env[65522]: _type = "Task" [ 771.508861] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.536554] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113920, 'name': CloneVM_Task} progress is 12%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.548090] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] 4238d766-2a5d-40d7-b2f6-5cb9a5354760/4238d766-2a5d-40d7-b2f6-5cb9a5354760.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 771.552765] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-948ef6a5-8676-434c-acc4-317f3e5791f9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.575386] env[65522]: DEBUG nova.compute.manager [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 771.577939] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113917, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517709} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.584391] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] bb20d99f-fc49-47bf-a6b7-74ea966b71eb/bb20d99f-fc49-47bf-a6b7-74ea966b71eb.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 771.584391] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 771.584670] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113921, 'name': CreateVM_Task} progress is 15%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.585477] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0e217e17-8912-4271-a47c-aeacd5d3299d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.594808] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113918, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.597255] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 771.597255] env[65522]: value = "task-5113922" [ 771.597255] env[65522]: _type = "Task" [ 771.597255] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.604876] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 771.604876] env[65522]: value = "task-5113923" [ 771.604876] env[65522]: _type = "Task" [ 771.604876] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.617827] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113922, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.626955] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113923, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.630676] env[65522]: DEBUG nova.virt.hardware [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 771.631300] env[65522]: DEBUG nova.virt.hardware [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 771.631300] env[65522]: DEBUG nova.virt.hardware [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 771.631586] env[65522]: DEBUG nova.virt.hardware [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 771.631829] env[65522]: DEBUG nova.virt.hardware [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 771.632106] env[65522]: DEBUG nova.virt.hardware [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 771.632469] env[65522]: DEBUG nova.virt.hardware [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.632714] env[65522]: DEBUG nova.virt.hardware [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 771.633079] env[65522]: DEBUG nova.virt.hardware [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 771.634066] env[65522]: DEBUG nova.virt.hardware [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 771.634066] env[65522]: DEBUG nova.virt.hardware [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 771.634972] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f38af6-04fb-48a5-aedd-11c6fc2db9c7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.647293] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6abea83-89f8-487c-811d-51b7cf177c7a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.657221] env[65522]: WARNING neutronclient.v2_0.client [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.657681] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.657864] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 771.657864] env[65522]: DEBUG nova.network.neutron [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 771.658080] env[65522]: DEBUG nova.objects.instance [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lazy-loading 'info_cache' on Instance uuid d5e8dd05-dc3c-4831-b4b0-ac100360f3e7 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 771.738863] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquiring lock "refresh_cache-6e10e846-cdb7-458e-b511-18ab1742228e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.739247] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquired lock "refresh_cache-6e10e846-cdb7-458e-b511-18ab1742228e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 771.739451] env[65522]: DEBUG nova.network.neutron [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 771.896459] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113919, 'name': ReconfigVM_Task, 'duration_secs': 0.467519} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.896959] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 1e8088ef-6354-40a0-95ee-06732e7f088f/1e8088ef-6354-40a0-95ee-06732e7f088f.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 771.897860] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-584c71d6-aa62-4406-855b-04131407f32f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.906469] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 771.906469] env[65522]: value = "task-5113924" [ 771.906469] env[65522]: _type = "Task" [ 771.906469] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.916735] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113924, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.959522] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.448s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 771.962135] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.862s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 771.962395] env[65522]: DEBUG nova.objects.instance [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lazy-loading 'resources' on Instance uuid f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 771.998323] env[65522]: INFO nova.scheduler.client.report [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Deleted allocations for instance 6abf58ca-ad6b-4fe8-844a-ca33e62a844d [ 772.016279] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113920, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.040659] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113918, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.041016] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113921, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.113556] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113922, 'name': ReconfigVM_Task, 'duration_secs': 0.41566} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.114445] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Reconfigured VM instance instance-00000027 to attach disk [datastore2] 4238d766-2a5d-40d7-b2f6-5cb9a5354760/4238d766-2a5d-40d7-b2f6-5cb9a5354760.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 772.115190] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e09c8422-11d7-49f6-ba3d-5f4557ddaa35 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.120775] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113923, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.480765} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.121859] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 772.122646] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717f8737-ae61-4673-a525-505a52c5d0d1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.127287] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 772.127287] env[65522]: value = "task-5113925" [ 772.127287] env[65522]: _type = "Task" [ 772.127287] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.152645] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] bb20d99f-fc49-47bf-a6b7-74ea966b71eb/bb20d99f-fc49-47bf-a6b7-74ea966b71eb.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 772.154398] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d08e1ba8-8c2b-4ecb-9f01-a9e95f230979 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.175710] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113925, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.183355] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 772.183355] env[65522]: value = "task-5113926" [ 772.183355] env[65522]: _type = "Task" [ 772.183355] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.196574] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113926, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.243791] env[65522]: WARNING neutronclient.v2_0.client [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 772.244553] env[65522]: WARNING openstack [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 772.244906] env[65522]: WARNING openstack [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 772.304599] env[65522]: DEBUG nova.network.neutron [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Successfully updated port: 9446c1bb-b443-4a52-879d-5fa6cb83f0dc {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 772.382527] env[65522]: DEBUG nova.network.neutron [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 772.421355] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113924, 'name': Rename_Task, 'duration_secs': 0.202617} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.421557] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 772.421809] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a52d5f1f-2578-46b6-9e4c-223b8fd3c67e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.429939] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 772.429939] env[65522]: value = "task-5113927" [ 772.429939] env[65522]: _type = "Task" [ 772.429939] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.439461] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113927, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.515030] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113920, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.515516] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91954964-39d9-4d8e-89ea-146f50ec698f tempest-AttachInterfacesUnderV243Test-1774195526 tempest-AttachInterfacesUnderV243Test-1774195526-project-member] Lock "6abf58ca-ad6b-4fe8-844a-ca33e62a844d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.378s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 772.532558] env[65522]: DEBUG oslo_vmware.api [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113918, 'name': PowerOnVM_Task, 'duration_secs': 1.027131} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.536757] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 772.537068] env[65522]: INFO nova.compute.manager [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Took 14.00 seconds to spawn the instance on the hypervisor. [ 772.537296] env[65522]: DEBUG nova.compute.manager [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 772.537726] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113921, 'name': CreateVM_Task, 'duration_secs': 0.691895} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.542089] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5087fa-0810-43bb-a2be-e3fb2fe6c05b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.545353] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 772.546610] env[65522]: WARNING neutronclient.v2_0.client [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 772.546966] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.547149] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 772.547620] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 772.548501] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-655230ec-3886-4374-9982-3fa19c7478c0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.564419] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 772.564419] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52610917-7859-3805-00fd-b829c8f821f5" [ 772.564419] env[65522]: _type = "Task" [ 772.564419] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.576101] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52610917-7859-3805-00fd-b829c8f821f5, 'name': SearchDatastore_Task, 'duration_secs': 0.01243} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.576450] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 772.576691] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.577084] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.577248] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 772.577431] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.577718] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-555d045a-1ea8-444f-b4de-97401685a788 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.588687] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 772.589023] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 772.589832] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0807d922-75c5-4145-bbb4-cb9d425e0305 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.597099] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 772.597099] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523ada2b-5b10-f369-8d4d-be8095d05b6a" [ 772.597099] env[65522]: _type = "Task" [ 772.597099] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.611122] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523ada2b-5b10-f369-8d4d-be8095d05b6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.615108] env[65522]: DEBUG nova.network.neutron [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 772.654327] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113925, 'name': Rename_Task, 'duration_secs': 0.171371} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.654327] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 772.654935] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf7671eb-d056-4f37-aad7-028862991321 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.663845] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 772.663845] env[65522]: value = "task-5113928" [ 772.663845] env[65522]: _type = "Task" [ 772.663845] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.676367] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113928, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.676920] env[65522]: WARNING neutronclient.v2_0.client [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 772.677581] env[65522]: WARNING openstack [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 772.677949] env[65522]: WARNING openstack [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 772.709900] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.807042] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.807042] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 772.807042] env[65522]: DEBUG nova.network.neutron [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 772.939634] env[65522]: WARNING openstack [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 772.941088] env[65522]: WARNING openstack [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 772.965475] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113927, 'name': PowerOnVM_Task, 'duration_secs': 0.520431} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.966157] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 772.967036] env[65522]: INFO nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Took 9.56 seconds to spawn the instance on the hypervisor. [ 772.967036] env[65522]: DEBUG nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 772.971026] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-813f3228-c081-457d-82ad-ef00061522cd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.016769] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113920, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.065009] env[65522]: WARNING neutronclient.v2_0.client [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 773.068153] env[65522]: WARNING openstack [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 773.068153] env[65522]: WARNING openstack [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 773.087295] env[65522]: INFO nova.compute.manager [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Took 33.57 seconds to build instance. [ 773.121520] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Releasing lock "refresh_cache-6e10e846-cdb7-458e-b511-18ab1742228e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 773.121520] env[65522]: DEBUG nova.compute.manager [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 773.121520] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 773.121920] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523ada2b-5b10-f369-8d4d-be8095d05b6a, 'name': SearchDatastore_Task, 'duration_secs': 0.012146} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.122824] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635d66f3-51e8-4a8b-ab75-f6cac00cb42d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.126672] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5819779b-8980-4c5f-9c7d-668e9d58088d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.133645] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 773.133645] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52275503-7c8f-5935-70ad-5dcc4a8525a8" [ 773.133645] env[65522]: _type = "Task" [ 773.133645] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.137034] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 773.142882] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74e0d152-dffd-4624-9fd5-fcfc0e8bdc47 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.154404] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52275503-7c8f-5935-70ad-5dcc4a8525a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.156482] env[65522]: DEBUG oslo_vmware.api [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 773.156482] env[65522]: value = "task-5113929" [ 773.156482] env[65522]: _type = "Task" [ 773.156482] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.171580] env[65522]: DEBUG oslo_vmware.api [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113929, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.178723] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113928, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.208713] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113926, 'name': ReconfigVM_Task, 'duration_secs': 0.811075} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.209130] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Reconfigured VM instance instance-00000029 to attach disk [datastore2] bb20d99f-fc49-47bf-a6b7-74ea966b71eb/bb20d99f-fc49-47bf-a6b7-74ea966b71eb.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 773.209838] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e14c7c4-c94f-4e21-8e22-2334138d4d56 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.217993] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 773.217993] env[65522]: value = "task-5113930" [ 773.217993] env[65522]: _type = "Task" [ 773.217993] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.229558] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113930, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.236770] env[65522]: DEBUG nova.objects.instance [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lazy-loading 'flavor' on Instance uuid 7c6f8218-602d-44f3-8012-de5a96972785 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 773.253567] env[65522]: DEBUG nova.network.neutron [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating instance_info_cache with network_info: [{"id": "fb1cd05a-757b-469c-9749-fd1ca4fcb168", "address": "fa:16:3e:b1:9d:ad", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.96", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb1cd05a-75", "ovs_interfaceid": "fb1cd05a-757b-469c-9749-fd1ca4fcb168", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 773.310152] env[65522]: WARNING openstack [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 773.310603] env[65522]: WARNING openstack [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 773.346740] env[65522]: DEBUG nova.compute.manager [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Received event network-vif-deleted-847d52c1-cfc5-4051-9969-5b8b567f330b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 773.346740] env[65522]: DEBUG nova.compute.manager [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Received event network-vif-plugged-03d0a08a-a560-4f8a-8cc9-24211dff931b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 773.346740] env[65522]: DEBUG oslo_concurrency.lockutils [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Acquiring lock "ea5b8e29-ebe6-444b-bd2d-029201003ee6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.346740] env[65522]: DEBUG oslo_concurrency.lockutils [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Lock "ea5b8e29-ebe6-444b-bd2d-029201003ee6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.346740] env[65522]: DEBUG oslo_concurrency.lockutils [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Lock "ea5b8e29-ebe6-444b-bd2d-029201003ee6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 773.346740] env[65522]: DEBUG nova.compute.manager [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] No waiting events found dispatching network-vif-plugged-03d0a08a-a560-4f8a-8cc9-24211dff931b {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 773.346740] env[65522]: WARNING nova.compute.manager [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Received unexpected event network-vif-plugged-03d0a08a-a560-4f8a-8cc9-24211dff931b for instance with vm_state active and task_state None. [ 773.346740] env[65522]: DEBUG nova.compute.manager [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Received event network-changed-03d0a08a-a560-4f8a-8cc9-24211dff931b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 773.347491] env[65522]: DEBUG nova.compute.manager [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Refreshing instance network info cache due to event network-changed-03d0a08a-a560-4f8a-8cc9-24211dff931b. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 773.347491] env[65522]: DEBUG oslo_concurrency.lockutils [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Acquiring lock "refresh_cache-ea5b8e29-ebe6-444b-bd2d-029201003ee6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.347491] env[65522]: DEBUG oslo_concurrency.lockutils [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Acquired lock "refresh_cache-ea5b8e29-ebe6-444b-bd2d-029201003ee6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 773.347491] env[65522]: DEBUG nova.network.neutron [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Refreshing network info cache for port 03d0a08a-a560-4f8a-8cc9-24211dff931b {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 773.377451] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13eb6af7-d773-4c1c-8e37-1621288a115d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.390408] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cdcc805-34f4-4b61-8468-a533c7430200 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.427675] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "f44db343-7e17-4780-9a67-245992ac1a63" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.427675] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "f44db343-7e17-4780-9a67-245992ac1a63" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.427896] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "f44db343-7e17-4780-9a67-245992ac1a63-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.427950] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "f44db343-7e17-4780-9a67-245992ac1a63-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.428139] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "f44db343-7e17-4780-9a67-245992ac1a63-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 773.432583] env[65522]: DEBUG nova.network.neutron [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 773.434571] env[65522]: INFO nova.compute.manager [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Terminating instance [ 773.440986] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d4e951e-797a-4f13-b441-aff89be63ec8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.444767] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquiring lock "9fa6e208-f913-4ecc-8548-ed6503fe6273" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.445016] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Lock "9fa6e208-f913-4ecc-8548-ed6503fe6273" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.455572] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac12dad4-7bd3-4412-8cac-952ee562212b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.474234] env[65522]: DEBUG nova.compute.provider_tree [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.495491] env[65522]: WARNING openstack [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 773.495868] env[65522]: WARNING openstack [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 773.516093] env[65522]: INFO nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Took 28.58 seconds to build instance. [ 773.517534] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Acquiring lock "cdee3a32-ad51-4364-8440-4744e8df322f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.517706] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Lock "cdee3a32-ad51-4364-8440-4744e8df322f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.524083] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113920, 'name': CloneVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.590763] env[65522]: DEBUG oslo_concurrency.lockutils [None req-75a21024-eeda-4926-bc8a-7edefd907bf1 tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Lock "ea5b8e29-ebe6-444b-bd2d-029201003ee6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.091s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 773.612205] env[65522]: WARNING neutronclient.v2_0.client [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 773.613074] env[65522]: WARNING openstack [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 773.613466] env[65522]: WARNING openstack [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 773.647829] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52275503-7c8f-5935-70ad-5dcc4a8525a8, 'name': SearchDatastore_Task, 'duration_secs': 0.033231} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.648031] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 773.648191] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 24f520ec-6a56-4f17-9ae4-2c856bf99582/24f520ec-6a56-4f17-9ae4-2c856bf99582.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 773.648728] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da3eb495-85ea-478f-bd0a-8a296bc266fd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.657208] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 773.657208] env[65522]: value = "task-5113931" [ 773.657208] env[65522]: _type = "Task" [ 773.657208] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.674421] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113931, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.674688] env[65522]: DEBUG oslo_vmware.api [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113929, 'name': PowerOffVM_Task, 'duration_secs': 0.210895} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.675379] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 773.675580] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 773.676020] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a9a9a655-5209-4e61-a069-65bcbdf33b73 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.683229] env[65522]: DEBUG oslo_vmware.api [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113928, 'name': PowerOnVM_Task, 'duration_secs': 0.527485} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.683954] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 773.684171] env[65522]: INFO nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Took 12.86 seconds to spawn the instance on the hypervisor. [ 773.684340] env[65522]: DEBUG nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 773.685206] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72de3731-390c-43b5-b4df-6d59fd17364d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.709409] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 773.709622] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 773.709897] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Deleting the datastore file [datastore2] 6e10e846-cdb7-458e-b511-18ab1742228e {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 773.710090] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a28b5f0-ff07-416f-a97b-5b9a49e73fd6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.717937] env[65522]: DEBUG oslo_vmware.api [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for the task: (returnval){ [ 773.717937] env[65522]: value = "task-5113933" [ 773.717937] env[65522]: _type = "Task" [ 773.717937] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.733664] env[65522]: DEBUG oslo_vmware.api [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113933, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.738721] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113930, 'name': Rename_Task, 'duration_secs': 0.210972} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.738772] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 773.739204] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-23cfda7c-3a77-4418-9bf0-985cef949bdb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.746458] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.746723] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 773.746976] env[65522]: DEBUG nova.network.neutron [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 773.747347] env[65522]: DEBUG nova.objects.instance [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lazy-loading 'info_cache' on Instance uuid 7c6f8218-602d-44f3-8012-de5a96972785 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 773.752483] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 773.752483] env[65522]: value = "task-5113934" [ 773.752483] env[65522]: _type = "Task" [ 773.752483] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.763803] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Releasing lock "refresh_cache-d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 773.766418] env[65522]: DEBUG nova.objects.instance [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lazy-loading 'migration_context' on Instance uuid d5e8dd05-dc3c-4831-b4b0-ac100360f3e7 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 773.779123] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113934, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.854404] env[65522]: WARNING neutronclient.v2_0.client [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 773.854714] env[65522]: WARNING openstack [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 773.854993] env[65522]: WARNING openstack [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 773.947317] env[65522]: DEBUG nova.compute.manager [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 773.947733] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 773.948289] env[65522]: DEBUG nova.compute.manager [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 773.952223] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184b83c2-e8e5-47a6-9a80-ded15470f87b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.964364] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 773.964783] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1217076b-07c7-469d-8695-4836694ac796 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.977185] env[65522]: DEBUG nova.scheduler.client.report [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 774.018546] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113920, 'name': CloneVM_Task, 'duration_secs': 2.083389} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.018873] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Created linked-clone VM from snapshot [ 774.019420] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "1e8088ef-6354-40a0-95ee-06732e7f088f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.099s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 774.020408] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe77366-a367-48bf-ba35-6673614302ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.032295] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Uploading image 2b17de9f-63a4-459f-8af1-578cd934c183 {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 774.043129] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 774.044565] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 774.044818] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Deleting the datastore file [datastore1] f44db343-7e17-4780-9a67-245992ac1a63 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 774.045215] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d8cb3d0-c124-44d6-be29-95904b404901 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.057399] env[65522]: DEBUG oslo_vmware.api [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 774.057399] env[65522]: value = "task-5113936" [ 774.057399] env[65522]: _type = "Task" [ 774.057399] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.080900] env[65522]: DEBUG oslo_vmware.api [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113936, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.084522] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 774.084522] env[65522]: value = "vm-994776" [ 774.084522] env[65522]: _type = "VirtualMachine" [ 774.084522] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 774.085041] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-2c60d0b3-08cf-4da4-a191-1dddc746c0a1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.094920] env[65522]: DEBUG nova.compute.manager [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 774.100107] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lease: (returnval){ [ 774.100107] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527a00de-a50d-e03a-75b0-76996c8e3cf2" [ 774.100107] env[65522]: _type = "HttpNfcLease" [ 774.100107] env[65522]: } obtained for exporting VM: (result){ [ 774.100107] env[65522]: value = "vm-994776" [ 774.100107] env[65522]: _type = "VirtualMachine" [ 774.100107] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 774.100451] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the lease: (returnval){ [ 774.100451] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527a00de-a50d-e03a-75b0-76996c8e3cf2" [ 774.100451] env[65522]: _type = "HttpNfcLease" [ 774.100451] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 774.116914] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 774.116914] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527a00de-a50d-e03a-75b0-76996c8e3cf2" [ 774.116914] env[65522]: _type = "HttpNfcLease" [ 774.116914] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 774.174824] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113931, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.205315] env[65522]: INFO nova.compute.manager [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Took 29.32 seconds to build instance. [ 774.236439] env[65522]: DEBUG oslo_vmware.api [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Task: {'id': task-5113933, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159152} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.236718] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 774.236902] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 774.237222] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 774.237350] env[65522]: INFO nova.compute.manager [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 774.237693] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 774.237937] env[65522]: DEBUG nova.compute.manager [-] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 774.238072] env[65522]: DEBUG nova.network.neutron [-] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 774.238390] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.239044] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.239339] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.250961] env[65522]: DEBUG nova.objects.base [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Object Instance<7c6f8218-602d-44f3-8012-de5a96972785> lazy-loaded attributes: flavor,info_cache {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 774.269883] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113934, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.273466] env[65522]: DEBUG nova.objects.base [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 774.274994] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f370a5e-b95e-49a9-984a-acea8bc1cdaa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.306962] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-328da8d7-8bd5-45d8-8553-3509976984b7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.314393] env[65522]: DEBUG oslo_vmware.api [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 774.314393] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f0081f-a399-f46a-2c7b-8ec25dfea6cb" [ 774.314393] env[65522]: _type = "Task" [ 774.314393] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.325640] env[65522]: DEBUG oslo_vmware.api [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f0081f-a399-f46a-2c7b-8ec25dfea6cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.326563] env[65522]: DEBUG nova.network.neutron [-] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 774.326816] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.346537] env[65522]: DEBUG nova.network.neutron [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating instance_info_cache with network_info: [{"id": "9446c1bb-b443-4a52-879d-5fa6cb83f0dc", "address": "fa:16:3e:54:30:7c", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9446c1bb-b4", "ovs_interfaceid": "9446c1bb-b443-4a52-879d-5fa6cb83f0dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 774.467560] env[65522]: WARNING openstack [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.468691] env[65522]: WARNING openstack [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.483736] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.522s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 774.487058] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.544s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 774.487058] env[65522]: DEBUG nova.objects.instance [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Lazy-loading 'resources' on Instance uuid 20855931-a02d-4b53-b998-3d89d4ff5d1a {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 774.489428] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 774.532936] env[65522]: INFO nova.scheduler.client.report [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleted allocations for instance f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7 [ 774.559274] env[65522]: WARNING neutronclient.v2_0.client [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.559274] env[65522]: WARNING openstack [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.559274] env[65522]: WARNING openstack [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.590428] env[65522]: DEBUG oslo_vmware.api [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113936, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.620151] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 774.620151] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527a00de-a50d-e03a-75b0-76996c8e3cf2" [ 774.620151] env[65522]: _type = "HttpNfcLease" [ 774.620151] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 774.620151] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 774.620151] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527a00de-a50d-e03a-75b0-76996c8e3cf2" [ 774.620151] env[65522]: _type = "HttpNfcLease" [ 774.620151] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 774.620952] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1926adef-50b1-46ad-90e3-2e5c74280ca7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.438553] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.445585] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a6cdd5a-55d2-4469-9766-102cffb81d5d tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "4238d766-2a5d-40d7-b2f6-5cb9a5354760" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.575s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 775.445995] env[65522]: WARNING neutronclient.v2_0.client [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 775.446911] env[65522]: WARNING openstack [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 775.446911] env[65522]: WARNING openstack [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 775.455442] env[65522]: DEBUG nova.network.neutron [-] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 775.456906] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 775.457292] env[65522]: DEBUG nova.compute.manager [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Instance network_info: |[{"id": "9446c1bb-b443-4a52-879d-5fa6cb83f0dc", "address": "fa:16:3e:54:30:7c", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9446c1bb-b4", "ovs_interfaceid": "9446c1bb-b443-4a52-879d-5fa6cb83f0dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 775.463464] env[65522]: DEBUG nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Received event network-changed-38d3255d-8d8f-4b7f-927e-5a7330420b08 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 775.464490] env[65522]: DEBUG nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Refreshing instance network info cache due to event network-changed-38d3255d-8d8f-4b7f-927e-5a7330420b08. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 775.464490] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Acquiring lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.464490] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Acquired lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 775.464490] env[65522]: DEBUG nova.network.neutron [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Refreshing network info cache for port 38d3255d-8d8f-4b7f-927e-5a7330420b08 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 775.480724] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:30:7c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6fb0104-186b-4288-b87e-634893f46f01', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9446c1bb-b443-4a52-879d-5fa6cb83f0dc', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 775.489649] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 775.494827] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Acquiring lock "1be995d0-3141-4dcd-ad9a-278d9874e457" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.494827] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Lock "1be995d0-3141-4dcd-ad9a-278d9874e457" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 775.499349] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 775.504798] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-071c2968-492f-4bad-b224-305bb5327328 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.521201] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a96318-8dd3-14da-0ac1-e58f23083277/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 775.521399] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a96318-8dd3-14da-0ac1-e58f23083277/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 775.529327] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5d944f72-36fa-4283-88af-378f438be2e0 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.721s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 775.537086] env[65522]: DEBUG oslo_vmware.api [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113934, 'name': PowerOnVM_Task, 'duration_secs': 0.686502} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.538344] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113931, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681855} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.538344] env[65522]: DEBUG oslo_vmware.api [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f0081f-a399-f46a-2c7b-8ec25dfea6cb, 'name': SearchDatastore_Task, 'duration_secs': 0.019279} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.542699] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 775.542931] env[65522]: INFO nova.compute.manager [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Took 9.42 seconds to spawn the instance on the hypervisor. [ 775.543139] env[65522]: DEBUG nova.compute.manager [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 775.543465] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 24f520ec-6a56-4f17-9ae4-2c856bf99582/24f520ec-6a56-4f17-9ae4-2c856bf99582.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 775.543662] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 775.543973] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.544683] env[65522]: DEBUG oslo_vmware.api [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5113936, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.548825} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.600336] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13e0153-a2e9-44a6-9edf-fb9ac49ae807 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.603286] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b10c19de-a88b-416e-8131-724ecd88da3a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.605907] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 775.606075] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 775.606257] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 775.606443] env[65522]: INFO nova.compute.manager [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Took 1.66 seconds to destroy the instance on the hypervisor. [ 775.606686] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 775.613040] env[65522]: DEBUG nova.compute.manager [-] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 775.613175] env[65522]: DEBUG nova.network.neutron [-] [instance: f44db343-7e17-4780-9a67-245992ac1a63] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 775.613442] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 775.614033] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 775.614327] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 775.621576] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 775.621576] env[65522]: value = "task-5113938" [ 775.621576] env[65522]: _type = "Task" [ 775.621576] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.633632] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 775.633632] env[65522]: value = "task-5113939" [ 775.633632] env[65522]: _type = "Task" [ 775.633632] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.644891] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113938, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.652191] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113939, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.667157] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8a667d55-c769-42dd-81f6-4d88c1621b7f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.854613] env[65522]: DEBUG nova.network.neutron [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Updated VIF entry in instance network info cache for port 03d0a08a-a560-4f8a-8cc9-24211dff931b. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 775.855540] env[65522]: DEBUG nova.network.neutron [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Updating instance_info_cache with network_info: [{"id": "03d0a08a-a560-4f8a-8cc9-24211dff931b", "address": "fa:16:3e:d8:09:37", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03d0a08a-a5", "ovs_interfaceid": "03d0a08a-a560-4f8a-8cc9-24211dff931b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 775.970025] env[65522]: INFO nova.compute.manager [-] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Took 1.73 seconds to deallocate network for instance. [ 775.970948] env[65522]: WARNING neutronclient.v2_0.client [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 775.971762] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 775.972132] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 776.007078] env[65522]: DEBUG nova.compute.manager [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 776.063294] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 776.134164] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113938, 'name': CreateVM_Task, 'duration_secs': 0.453433} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.134267] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 776.135033] env[65522]: WARNING neutronclient.v2_0.client [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 776.135638] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.135876] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 776.136343] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 776.144043] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc875fca-0c92-4233-b67b-197bfb227c3e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.158996] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113939, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082314} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.159310] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 776.159310] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528b5b61-25c9-a5ed-3342-dda6a2323723" [ 776.159310] env[65522]: _type = "Task" [ 776.159310] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.160022] env[65522]: INFO nova.compute.manager [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Took 29.10 seconds to build instance. [ 776.161947] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 776.166976] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce84b53b-6fbc-4fdb-9454-3d48edbc1373 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.183917] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528b5b61-25c9-a5ed-3342-dda6a2323723, 'name': SearchDatastore_Task, 'duration_secs': 0.018763} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.195154] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 776.195613] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 776.195862] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.196109] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 776.196574] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 776.206267] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] 24f520ec-6a56-4f17-9ae4-2c856bf99582/24f520ec-6a56-4f17-9ae4-2c856bf99582.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 776.210073] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-60010a89-9ef8-44f8-abc0-ec1e139b5465 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.213023] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad312b5f-d89f-403a-aa86-b92de96586d5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.241448] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 776.241448] env[65522]: value = "task-5113940" [ 776.241448] env[65522]: _type = "Task" [ 776.241448] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.246106] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 776.246461] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 776.252856] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4eaa6d89-fe71-4e2b-963f-92e83660376d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.266666] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113940, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.266666] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 776.266666] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52eeeaa4-d585-3142-4b54-94a775b34a2f" [ 776.266666] env[65522]: _type = "Task" [ 776.266666] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.282039] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52eeeaa4-d585-3142-4b54-94a775b34a2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.357083] env[65522]: DEBUG nova.compute.manager [None req-1400d501-a2f0-4fa9-bb0a-8dcd6e3bb8b3 tempest-ServerDiagnosticsTest-1218861246 tempest-ServerDiagnosticsTest-1218861246-project-admin] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 776.361750] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691fcbef-fb08-4c58-b6ce-d2f27342d767 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.369442] env[65522]: DEBUG oslo_concurrency.lockutils [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Releasing lock "refresh_cache-ea5b8e29-ebe6-444b-bd2d-029201003ee6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 776.369813] env[65522]: DEBUG nova.compute.manager [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Received event network-vif-plugged-652babe2-668c-4364-92c2-8ceaba23c0ac {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 776.370147] env[65522]: DEBUG oslo_concurrency.lockutils [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Acquiring lock "bb20d99f-fc49-47bf-a6b7-74ea966b71eb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 776.370462] env[65522]: DEBUG oslo_concurrency.lockutils [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Lock "bb20d99f-fc49-47bf-a6b7-74ea966b71eb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 776.370558] env[65522]: DEBUG oslo_concurrency.lockutils [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Lock "bb20d99f-fc49-47bf-a6b7-74ea966b71eb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 776.370741] env[65522]: DEBUG nova.compute.manager [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] No waiting events found dispatching network-vif-plugged-652babe2-668c-4364-92c2-8ceaba23c0ac {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 776.370906] env[65522]: WARNING nova.compute.manager [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Received unexpected event network-vif-plugged-652babe2-668c-4364-92c2-8ceaba23c0ac for instance with vm_state building and task_state spawning. [ 776.371081] env[65522]: DEBUG nova.compute.manager [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Received event network-changed-652babe2-668c-4364-92c2-8ceaba23c0ac {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 776.371368] env[65522]: DEBUG nova.compute.manager [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Refreshing instance network info cache due to event network-changed-652babe2-668c-4364-92c2-8ceaba23c0ac. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 776.371560] env[65522]: DEBUG oslo_concurrency.lockutils [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Acquiring lock "refresh_cache-bb20d99f-fc49-47bf-a6b7-74ea966b71eb" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.371696] env[65522]: DEBUG oslo_concurrency.lockutils [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Acquired lock "refresh_cache-bb20d99f-fc49-47bf-a6b7-74ea966b71eb" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 776.372088] env[65522]: DEBUG nova.network.neutron [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Refreshing network info cache for port 652babe2-668c-4364-92c2-8ceaba23c0ac {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 776.378686] env[65522]: INFO nova.compute.manager [None req-1400d501-a2f0-4fa9-bb0a-8dcd6e3bb8b3 tempest-ServerDiagnosticsTest-1218861246 tempest-ServerDiagnosticsTest-1218861246-project-admin] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Retrieving diagnostics [ 776.379571] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-906dd677-fbbb-47c1-a515-874d6d9ba477 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.443926] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33730e38-1353-4eeb-a01a-b05f143a642e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.450288] env[65522]: WARNING openstack [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 776.451030] env[65522]: WARNING openstack [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 776.469036] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25315f32-07bf-4e0e-b587-02f78f9017b3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.509797] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 776.518932] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3abec51c-cb83-4873-9659-13142044dec6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.531488] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b04ed1b-6b5e-437f-9ae8-aa422fea31dc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.552311] env[65522]: DEBUG nova.compute.provider_tree [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.557280] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 776.618861] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 776.619321] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 776.667563] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9655fd5a-32ac-4599-bc10-99438a231d19 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "bb20d99f-fc49-47bf-a6b7-74ea966b71eb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.634s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 776.758013] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113940, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.777427] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52eeeaa4-d585-3142-4b54-94a775b34a2f, 'name': SearchDatastore_Task, 'duration_secs': 0.019149} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.778780] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60202a36-5e08-414e-97d1-8fc6a7c41cf4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.787053] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 776.787053] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5247f2d4-4b8f-d7f9-c473-06596c114ea2" [ 776.787053] env[65522]: _type = "Task" [ 776.787053] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.800256] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5247f2d4-4b8f-d7f9-c473-06596c114ea2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.801921] env[65522]: WARNING neutronclient.v2_0.client [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 776.802812] env[65522]: WARNING openstack [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 776.803040] env[65522]: WARNING openstack [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 776.875260] env[65522]: WARNING neutronclient.v2_0.client [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 776.875963] env[65522]: WARNING openstack [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 776.876439] env[65522]: WARNING openstack [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 776.963897] env[65522]: WARNING neutronclient.v2_0.client [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 776.964659] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 776.965010] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 777.013721] env[65522]: DEBUG nova.network.neutron [-] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 777.072919] env[65522]: DEBUG nova.scheduler.client.report [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 777.104861] env[65522]: DEBUG nova.network.neutron [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance_info_cache with network_info: [{"id": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "address": "fa:16:3e:69:05:d7", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf16c93a-75", "ovs_interfaceid": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 777.258758] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113940, 'name': ReconfigVM_Task, 'duration_secs': 0.547613} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.259312] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Reconfigured VM instance instance-0000002a to attach disk [datastore2] 24f520ec-6a56-4f17-9ae4-2c856bf99582/24f520ec-6a56-4f17-9ae4-2c856bf99582.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 777.260231] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2468b777-3411-454e-a0cd-529da33512db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.271119] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 777.271119] env[65522]: value = "task-5113941" [ 777.271119] env[65522]: _type = "Task" [ 777.271119] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.273621] env[65522]: DEBUG nova.network.neutron [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Updated VIF entry in instance network info cache for port 38d3255d-8d8f-4b7f-927e-5a7330420b08. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 777.274257] env[65522]: DEBUG nova.network.neutron [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Updating instance_info_cache with network_info: [{"id": "d24163e8-4f08-4908-9307-c4edb0d4ffb7", "address": "fa:16:3e:e2:90:31", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd24163e8-4f", "ovs_interfaceid": "d24163e8-4f08-4908-9307-c4edb0d4ffb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "38d3255d-8d8f-4b7f-927e-5a7330420b08", "address": "fa:16:3e:76:d8:87", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap38d3255d-8d", "ovs_interfaceid": "38d3255d-8d8f-4b7f-927e-5a7330420b08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 777.287715] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113941, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.301608] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5247f2d4-4b8f-d7f9-c473-06596c114ea2, 'name': SearchDatastore_Task, 'duration_secs': 0.016058} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.302263] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 777.302636] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 495fa98f-a9db-4214-87cc-d29209d3cb62/495fa98f-a9db-4214-87cc-d29209d3cb62.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 777.302977] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94e9dbe8-215d-4edc-8e10-b42fba38db90 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.313049] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 777.313049] env[65522]: value = "task-5113942" [ 777.313049] env[65522]: _type = "Task" [ 777.313049] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.319511] env[65522]: WARNING openstack [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 777.320144] env[65522]: WARNING openstack [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 777.338795] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113942, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.515416] env[65522]: WARNING neutronclient.v2_0.client [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 777.516286] env[65522]: WARNING openstack [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 777.516286] env[65522]: WARNING openstack [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 777.535036] env[65522]: INFO nova.compute.manager [-] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Took 1.92 seconds to deallocate network for instance. [ 777.581030] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.094s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 777.586382] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.557s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 777.586382] env[65522]: INFO nova.compute.claims [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.609266] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 777.622066] env[65522]: INFO nova.scheduler.client.report [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Deleted allocations for instance 20855931-a02d-4b53-b998-3d89d4ff5d1a [ 777.701036] env[65522]: DEBUG nova.compute.manager [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Received event network-vif-plugged-9446c1bb-b443-4a52-879d-5fa6cb83f0dc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 777.701036] env[65522]: DEBUG oslo_concurrency.lockutils [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] Acquiring lock "495fa98f-a9db-4214-87cc-d29209d3cb62-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 777.705064] env[65522]: DEBUG oslo_concurrency.lockutils [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] Lock "495fa98f-a9db-4214-87cc-d29209d3cb62-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 777.705064] env[65522]: DEBUG oslo_concurrency.lockutils [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] Lock "495fa98f-a9db-4214-87cc-d29209d3cb62-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 777.705064] env[65522]: DEBUG nova.compute.manager [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] No waiting events found dispatching network-vif-plugged-9446c1bb-b443-4a52-879d-5fa6cb83f0dc {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 777.705064] env[65522]: WARNING nova.compute.manager [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Received unexpected event network-vif-plugged-9446c1bb-b443-4a52-879d-5fa6cb83f0dc for instance with vm_state building and task_state spawning. [ 777.705064] env[65522]: DEBUG nova.compute.manager [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Received event network-changed-9446c1bb-b443-4a52-879d-5fa6cb83f0dc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 777.705064] env[65522]: DEBUG nova.compute.manager [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Refreshing instance network info cache due to event network-changed-9446c1bb-b443-4a52-879d-5fa6cb83f0dc. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 777.705064] env[65522]: DEBUG oslo_concurrency.lockutils [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] Acquiring lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.705064] env[65522]: DEBUG oslo_concurrency.lockutils [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] Acquired lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 777.705064] env[65522]: DEBUG nova.network.neutron [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Refreshing network info cache for port 9446c1bb-b443-4a52-879d-5fa6cb83f0dc {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 777.715878] env[65522]: DEBUG nova.network.neutron [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Updated VIF entry in instance network info cache for port 652babe2-668c-4364-92c2-8ceaba23c0ac. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 777.716515] env[65522]: DEBUG nova.network.neutron [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Updating instance_info_cache with network_info: [{"id": "652babe2-668c-4364-92c2-8ceaba23c0ac", "address": "fa:16:3e:fe:e2:bf", "network": {"id": "dc9d8462-3539-4cc5-b21a-e1136a8df8e6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1580876213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "b5c14a856d454048856539895f220c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap652babe2-66", "ovs_interfaceid": "652babe2-668c-4364-92c2-8ceaba23c0ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 777.787222] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Releasing lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 777.789456] env[65522]: DEBUG nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Received event network-vif-plugged-cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 777.789456] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Acquiring lock "4238d766-2a5d-40d7-b2f6-5cb9a5354760-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 777.789456] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Lock "4238d766-2a5d-40d7-b2f6-5cb9a5354760-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 777.790023] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Lock "4238d766-2a5d-40d7-b2f6-5cb9a5354760-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 777.790148] env[65522]: DEBUG nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] No waiting events found dispatching network-vif-plugged-cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 777.790353] env[65522]: WARNING nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Received unexpected event network-vif-plugged-cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c for instance with vm_state active and task_state None. [ 777.790525] env[65522]: DEBUG nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Received event network-changed-cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 777.790674] env[65522]: DEBUG nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Refreshing instance network info cache due to event network-changed-cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 777.790884] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Acquiring lock "refresh_cache-4238d766-2a5d-40d7-b2f6-5cb9a5354760" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.791014] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Acquired lock "refresh_cache-4238d766-2a5d-40d7-b2f6-5cb9a5354760" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 777.791198] env[65522]: DEBUG nova.network.neutron [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Refreshing network info cache for port cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 777.804932] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113941, 'name': Rename_Task, 'duration_secs': 0.220485} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.804932] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 777.805198] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-79c5981f-c404-44bc-bfe2-1488849aee0f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.818396] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 777.818396] env[65522]: value = "task-5113943" [ 777.818396] env[65522]: _type = "Task" [ 777.818396] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.838744] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113942, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.844399] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113943, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.905809] env[65522]: DEBUG oslo_concurrency.lockutils [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "0d52b207-ac69-48ce-b3af-7a83a499886e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 777.906076] env[65522]: DEBUG oslo_concurrency.lockutils [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "0d52b207-ac69-48ce-b3af-7a83a499886e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 777.906457] env[65522]: DEBUG oslo_concurrency.lockutils [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "0d52b207-ac69-48ce-b3af-7a83a499886e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 777.906796] env[65522]: DEBUG oslo_concurrency.lockutils [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "0d52b207-ac69-48ce-b3af-7a83a499886e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 777.907164] env[65522]: DEBUG oslo_concurrency.lockutils [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "0d52b207-ac69-48ce-b3af-7a83a499886e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 777.910506] env[65522]: INFO nova.compute.manager [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Terminating instance [ 778.044017] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 778.131876] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4f38ee73-3636-4cb9-9e59-549191f920c2 tempest-VolumesAdminNegativeTest-2117091976 tempest-VolumesAdminNegativeTest-2117091976-project-member] Lock "20855931-a02d-4b53-b998-3d89d4ff5d1a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.267s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 778.153583] env[65522]: DEBUG nova.compute.manager [req-5ef0286a-c6b8-4dd9-9f11-6fd990cc32de req-cd69a0c9-5ab6-4f15-9e29-0e462b058fdd service nova] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Received event network-vif-deleted-5e8b1f95-61cc-40fa-89f3-0581ba1fbe4e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 778.211542] env[65522]: WARNING neutronclient.v2_0.client [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 778.212309] env[65522]: WARNING openstack [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.212657] env[65522]: WARNING openstack [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.224634] env[65522]: DEBUG oslo_concurrency.lockutils [req-209ddd42-8ab9-4819-8731-e217f7b457b8 req-e5cf6f09-d780-4359-986d-22a18efbf7d1 service nova] Releasing lock "refresh_cache-bb20d99f-fc49-47bf-a6b7-74ea966b71eb" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 778.307665] env[65522]: WARNING neutronclient.v2_0.client [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 778.307665] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.307665] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.334588] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113942, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681157} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.335792] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 495fa98f-a9db-4214-87cc-d29209d3cb62/495fa98f-a9db-4214-87cc-d29209d3cb62.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 778.336066] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 778.336988] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e40903c3-ba49-4812-aae8-e870ed7ca521 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.342971] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113943, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.353471] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 778.353471] env[65522]: value = "task-5113944" [ 778.353471] env[65522]: _type = "Task" [ 778.353471] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.364830] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113944, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.415544] env[65522]: DEBUG nova.compute.manager [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 778.415873] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 778.417402] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533673c8-6595-42e3-95e1-9a4489d12847 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.427621] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 778.427621] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a0494add-de25-4940-bb1e-66e23a820518 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.437348] env[65522]: DEBUG oslo_vmware.api [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 778.437348] env[65522]: value = "task-5113945" [ 778.437348] env[65522]: _type = "Task" [ 778.437348] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.447833] env[65522]: DEBUG oslo_vmware.api [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113945, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.543762] env[65522]: WARNING openstack [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.545585] env[65522]: WARNING openstack [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.563177] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.563617] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.617250] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 778.617687] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-54d0d3b6-03cd-4c1b-b3db-e724e086df02 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.626397] env[65522]: DEBUG oslo_vmware.api [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 778.626397] env[65522]: value = "task-5113946" [ 778.626397] env[65522]: _type = "Task" [ 778.626397] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.646867] env[65522]: DEBUG oslo_vmware.api [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113946, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.742307] env[65522]: WARNING neutronclient.v2_0.client [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 778.742307] env[65522]: WARNING openstack [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.742307] env[65522]: WARNING openstack [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.764796] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Acquiring lock "ea5b8e29-ebe6-444b-bd2d-029201003ee6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 778.766626] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Lock "ea5b8e29-ebe6-444b-bd2d-029201003ee6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 778.766626] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Acquiring lock "ea5b8e29-ebe6-444b-bd2d-029201003ee6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 778.766626] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Lock "ea5b8e29-ebe6-444b-bd2d-029201003ee6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 778.766626] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Lock "ea5b8e29-ebe6-444b-bd2d-029201003ee6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 778.772733] env[65522]: WARNING neutronclient.v2_0.client [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 778.773233] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.773720] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.795694] env[65522]: INFO nova.compute.manager [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Terminating instance [ 778.838293] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113943, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.870944] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113944, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082005} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.875271] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 778.878808] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887154e8-daa1-40a3-9950-c787611ecb0f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.911964] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] 495fa98f-a9db-4214-87cc-d29209d3cb62/495fa98f-a9db-4214-87cc-d29209d3cb62.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 778.920353] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c846e9f-5c40-4926-81b5-dcc4d65b1552 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.947054] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 778.947054] env[65522]: value = "task-5113947" [ 778.947054] env[65522]: _type = "Task" [ 778.947054] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.954384] env[65522]: DEBUG oslo_vmware.api [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113945, 'name': PowerOffVM_Task, 'duration_secs': 0.347395} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.955328] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 778.955614] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 778.955954] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e4a85494-d1d3-4ab6-9b59-b4a7335bcee7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.962513] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113947, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.992295] env[65522]: DEBUG nova.network.neutron [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updated VIF entry in instance network info cache for port 9446c1bb-b443-4a52-879d-5fa6cb83f0dc. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 778.992733] env[65522]: DEBUG nova.network.neutron [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating instance_info_cache with network_info: [{"id": "9446c1bb-b443-4a52-879d-5fa6cb83f0dc", "address": "fa:16:3e:54:30:7c", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9446c1bb-b4", "ovs_interfaceid": "9446c1bb-b443-4a52-879d-5fa6cb83f0dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 779.038723] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 779.039568] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 779.039568] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Deleting the datastore file [datastore1] 0d52b207-ac69-48ce-b3af-7a83a499886e {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 779.039717] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b352a22-b923-416d-a0a5-9341d40d4ed3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.057024] env[65522]: DEBUG oslo_vmware.api [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 779.057024] env[65522]: value = "task-5113949" [ 779.057024] env[65522]: _type = "Task" [ 779.057024] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.072728] env[65522]: DEBUG oslo_vmware.api [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113949, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.103508] env[65522]: DEBUG nova.network.neutron [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Updated VIF entry in instance network info cache for port cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 779.103508] env[65522]: DEBUG nova.network.neutron [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Updating instance_info_cache with network_info: [{"id": "cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c", "address": "fa:16:3e:2c:05:d9", "network": {"id": "0ab23355-59a7-409b-89f3-089fbd03db5e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-960579569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acbde84c090e4c0c919728b809c73d02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf5b96ab-5c", "ovs_interfaceid": "cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 779.143986] env[65522]: DEBUG oslo_vmware.api [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113946, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.305038] env[65522]: DEBUG nova.compute.manager [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 779.305790] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 779.308364] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67635a9d-2113-4935-96a8-d5b823c1765e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.326058] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 779.326644] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3bbf2ca5-85fd-435b-bd44-18af0a79efa6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.342788] env[65522]: DEBUG oslo_vmware.api [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5113943, 'name': PowerOnVM_Task, 'duration_secs': 1.468046} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.348495] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 779.348495] env[65522]: INFO nova.compute.manager [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Took 10.48 seconds to spawn the instance on the hypervisor. [ 779.348495] env[65522]: DEBUG nova.compute.manager [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 779.348495] env[65522]: DEBUG oslo_vmware.api [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Waiting for the task: (returnval){ [ 779.348495] env[65522]: value = "task-5113950" [ 779.348495] env[65522]: _type = "Task" [ 779.348495] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.349415] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20033c51-d566-403c-ad47-9fde7235f492 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.362684] env[65522]: DEBUG oslo_vmware.api [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113950, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.471123] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113947, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.486180] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55be0aee-daca-4511-a792-eb467b093ca0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.495582] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5487cae1-7c4d-4e1f-be5c-ffe25cc6247b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.500018] env[65522]: DEBUG oslo_concurrency.lockutils [req-ec28322d-be85-425c-81be-521adcaa307a req-2d1dc411-6333-4d44-9257-45cb2c7b7b93 service nova] Releasing lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 779.541023] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231475d2-f803-4ba1-a0b9-8efb2ee760a6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.553926] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "4238d766-2a5d-40d7-b2f6-5cb9a5354760" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.554558] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "4238d766-2a5d-40d7-b2f6-5cb9a5354760" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.554558] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "4238d766-2a5d-40d7-b2f6-5cb9a5354760-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.554840] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "4238d766-2a5d-40d7-b2f6-5cb9a5354760-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.554840] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "4238d766-2a5d-40d7-b2f6-5cb9a5354760-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 779.557934] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9f491e-55b0-452f-bceb-f28df32000e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.565223] env[65522]: INFO nova.compute.manager [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Terminating instance [ 779.585532] env[65522]: DEBUG nova.compute.provider_tree [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.594614] env[65522]: DEBUG oslo_vmware.api [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5113949, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.335368} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.595229] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 779.595501] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 779.595639] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 779.597021] env[65522]: INFO nova.compute.manager [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Took 1.18 seconds to destroy the instance on the hypervisor. [ 779.597021] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 779.597021] env[65522]: DEBUG nova.compute.manager [-] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 779.597021] env[65522]: DEBUG nova.network.neutron [-] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 779.597021] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 779.597609] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 779.597872] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 779.610342] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Releasing lock "refresh_cache-4238d766-2a5d-40d7-b2f6-5cb9a5354760" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 779.610613] env[65522]: DEBUG nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Received event network-vif-plugged-3d5e8380-9410-49a2-873c-962835dabfaa {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 779.611814] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Acquiring lock "1e8088ef-6354-40a0-95ee-06732e7f088f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.611814] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Lock "1e8088ef-6354-40a0-95ee-06732e7f088f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.611814] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Lock "1e8088ef-6354-40a0-95ee-06732e7f088f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 779.611814] env[65522]: DEBUG nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] No waiting events found dispatching network-vif-plugged-3d5e8380-9410-49a2-873c-962835dabfaa {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 779.611814] env[65522]: WARNING nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Received unexpected event network-vif-plugged-3d5e8380-9410-49a2-873c-962835dabfaa for instance with vm_state active and task_state None. [ 779.611814] env[65522]: DEBUG nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Received event network-changed-3d5e8380-9410-49a2-873c-962835dabfaa {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 779.611814] env[65522]: DEBUG nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Refreshing instance network info cache due to event network-changed-3d5e8380-9410-49a2-873c-962835dabfaa. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 779.612406] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Acquiring lock "refresh_cache-1e8088ef-6354-40a0-95ee-06732e7f088f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.612406] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Acquired lock "refresh_cache-1e8088ef-6354-40a0-95ee-06732e7f088f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 779.612406] env[65522]: DEBUG nova.network.neutron [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Refreshing network info cache for port 3d5e8380-9410-49a2-873c-962835dabfaa {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 779.648394] env[65522]: DEBUG oslo_vmware.api [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5113946, 'name': PowerOnVM_Task, 'duration_secs': 0.759588} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.648394] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 779.648394] env[65522]: DEBUG nova.compute.manager [None req-e317c0e8-7998-460a-8226-9d8696a8e8d1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 779.650045] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc6c042-2ea5-4834-a62f-3207c9696ed3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.866590] env[65522]: DEBUG oslo_vmware.api [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113950, 'name': PowerOffVM_Task, 'duration_secs': 0.342717} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.866767] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 779.866928] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 779.867310] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-70187049-e41d-405c-a43f-ed8c78c16ee5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.882427] env[65522]: INFO nova.compute.manager [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Took 31.01 seconds to build instance. [ 779.971342] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113947, 'name': ReconfigVM_Task, 'duration_secs': 0.576359} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.971735] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 779.972013] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 779.972519] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Deleting the datastore file [datastore2] ea5b8e29-ebe6-444b-bd2d-029201003ee6 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 779.972688] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Reconfigured VM instance instance-0000002b to attach disk [datastore1] 495fa98f-a9db-4214-87cc-d29209d3cb62/495fa98f-a9db-4214-87cc-d29209d3cb62.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 779.975703] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4956b99a-2503-471b-acaa-8afda89fdf59 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.978227] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d204c31c-2272-4c01-b3e0-ad9836d79814 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.987741] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 779.987741] env[65522]: value = "task-5113953" [ 779.987741] env[65522]: _type = "Task" [ 779.987741] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.988370] env[65522]: DEBUG oslo_vmware.api [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Waiting for the task: (returnval){ [ 779.988370] env[65522]: value = "task-5113952" [ 779.988370] env[65522]: _type = "Task" [ 779.988370] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.003331] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113953, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.007376] env[65522]: DEBUG oslo_vmware.api [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113952, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.075051] env[65522]: DEBUG nova.compute.manager [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 780.075210] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.078042] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d4e917-ab05-4c8b-b201-d6499df8e281 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.087065] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 780.087432] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a470cce-8694-41f5-8d46-c984f6e90085 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.097459] env[65522]: DEBUG nova.scheduler.client.report [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 780.106307] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Acquiring lock "fdf56fa6-ad18-499e-a171-6242abb293bf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.107176] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Lock "fdf56fa6-ad18-499e-a171-6242abb293bf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.005s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.107517] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Acquiring lock "fdf56fa6-ad18-499e-a171-6242abb293bf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.107733] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Lock "fdf56fa6-ad18-499e-a171-6242abb293bf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.107906] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Lock "fdf56fa6-ad18-499e-a171-6242abb293bf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.110502] env[65522]: DEBUG oslo_vmware.api [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 780.110502] env[65522]: value = "task-5113954" [ 780.110502] env[65522]: _type = "Task" [ 780.110502] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.111381] env[65522]: INFO nova.compute.manager [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Terminating instance [ 780.121822] env[65522]: WARNING neutronclient.v2_0.client [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 780.124189] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 780.124189] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 780.140509] env[65522]: DEBUG oslo_vmware.api [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113954, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.178338] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "05319f69-2ac7-4649-a90a-3542d87a3e82" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.180240] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "05319f69-2ac7-4649-a90a-3542d87a3e82" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.183945] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 780.385060] env[65522]: DEBUG oslo_concurrency.lockutils [None req-74d70642-0a54-4bae-9ada-b295f64ee4bd tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "24f520ec-6a56-4f17-9ae4-2c856bf99582" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.524s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.502445] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113953, 'name': Rename_Task, 'duration_secs': 0.228335} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.503207] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 780.503507] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3134bac2-d868-439b-8a06-c6e2c4b09dbe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.508828] env[65522]: DEBUG oslo_vmware.api [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Task: {'id': task-5113952, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.221727} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.509678] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 780.509891] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 780.510105] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 780.510280] env[65522]: INFO nova.compute.manager [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Took 1.21 seconds to destroy the instance on the hypervisor. [ 780.510521] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 780.510715] env[65522]: DEBUG nova.compute.manager [-] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 780.510805] env[65522]: DEBUG nova.network.neutron [-] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 780.511194] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 780.511985] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 780.512393] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 780.522108] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 780.522108] env[65522]: value = "task-5113955" [ 780.522108] env[65522]: _type = "Task" [ 780.522108] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.534162] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113955, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.612313] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.028s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.612905] env[65522]: DEBUG nova.compute.manager [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 780.616042] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.396s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.616042] env[65522]: DEBUG nova.objects.instance [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Lazy-loading 'resources' on Instance uuid a09c12be-0aaa-4934-a1b3-5af29feba8a6 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 780.628331] env[65522]: DEBUG oslo_vmware.api [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113954, 'name': PowerOffVM_Task, 'duration_secs': 0.256146} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.629296] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 780.629470] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 780.629741] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3091bf3b-8759-4589-8216-a3e29fc20f71 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.632668] env[65522]: DEBUG nova.compute.manager [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 780.632879] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.634326] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecea4c92-c0c6-4905-bb18-e09bae986c75 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.643897] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 780.644269] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9efea369-9321-4c55-bfa9-1b5196627947 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.652975] env[65522]: DEBUG oslo_vmware.api [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Waiting for the task: (returnval){ [ 780.652975] env[65522]: value = "task-5113957" [ 780.652975] env[65522]: _type = "Task" [ 780.652975] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.672029] env[65522]: DEBUG oslo_vmware.api [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113957, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.687820] env[65522]: DEBUG nova.compute.manager [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 780.715915] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 780.716289] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 780.716657] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Deleting the datastore file [datastore2] 4238d766-2a5d-40d7-b2f6-5cb9a5354760 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 780.716925] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f3d9bbe2-d981-4dfe-b887-01620fee292c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.727743] env[65522]: DEBUG oslo_vmware.api [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 780.727743] env[65522]: value = "task-5113958" [ 780.727743] env[65522]: _type = "Task" [ 780.727743] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.738935] env[65522]: DEBUG oslo_vmware.api [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.038483] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113955, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.119939] env[65522]: DEBUG nova.compute.utils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 781.123463] env[65522]: DEBUG nova.compute.manager [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 781.126929] env[65522]: DEBUG nova.network.neutron [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 781.126929] env[65522]: WARNING neutronclient.v2_0.client [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.126929] env[65522]: WARNING neutronclient.v2_0.client [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.127260] env[65522]: WARNING openstack [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.129935] env[65522]: WARNING openstack [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.169322] env[65522]: DEBUG oslo_vmware.api [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113957, 'name': PowerOffVM_Task, 'duration_secs': 0.275339} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.169322] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 781.169322] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 781.169322] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9903466c-e824-4b00-a2d3-af8b185ea456 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.229341] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 781.245463] env[65522]: DEBUG oslo_vmware.api [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113958, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228858} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.248485] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 781.248762] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 781.249085] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 781.250590] env[65522]: INFO nova.compute.manager [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Took 1.17 seconds to destroy the instance on the hypervisor. [ 781.250590] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 781.250590] env[65522]: DEBUG nova.compute.manager [-] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 781.250590] env[65522]: DEBUG nova.network.neutron [-] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 781.250590] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.251339] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.251567] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.266988] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 781.267259] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 781.267443] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Deleting the datastore file [datastore1] fdf56fa6-ad18-499e-a171-6242abb293bf {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 781.272056] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2b8a6ec8-bd68-4f56-90a3-beba3c8bdfec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.280028] env[65522]: DEBUG oslo_vmware.api [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Waiting for the task: (returnval){ [ 781.280028] env[65522]: value = "task-5113960" [ 781.280028] env[65522]: _type = "Task" [ 781.280028] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.294757] env[65522]: DEBUG oslo_vmware.api [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113960, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.376479] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.540103] env[65522]: DEBUG oslo_vmware.api [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5113955, 'name': PowerOnVM_Task, 'duration_secs': 0.705756} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.540883] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 781.541017] env[65522]: INFO nova.compute.manager [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Took 9.97 seconds to spawn the instance on the hypervisor. [ 781.541204] env[65522]: DEBUG nova.compute.manager [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 781.541992] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c576223-8c17-44b0-8f7e-d35b891901d8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.624153] env[65522]: DEBUG nova.compute.manager [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 781.749983] env[65522]: DEBUG nova.network.neutron [-] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 781.793369] env[65522]: DEBUG oslo_vmware.api [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Task: {'id': task-5113960, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.296302} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.793369] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 781.793369] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 781.793869] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 781.793869] env[65522]: INFO nova.compute.manager [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Took 1.16 seconds to destroy the instance on the hypervisor. [ 781.793869] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 781.796791] env[65522]: DEBUG nova.compute.manager [-] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 781.796993] env[65522]: DEBUG nova.network.neutron [-] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 781.797711] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.797798] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.798069] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.811463] env[65522]: DEBUG nova.policy [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '792013a276474561bc625ac0e00cdb82', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2bd38f300424db78568e0b8d9982f27', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 781.815364] env[65522]: DEBUG oslo_concurrency.lockutils [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "interface-b2d70982-54e7-459c-a0d8-48bf4b6e4345-38d3255d-8d8f-4b7f-927e-5a7330420b08" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 781.816203] env[65522]: DEBUG oslo_concurrency.lockutils [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-b2d70982-54e7-459c-a0d8-48bf4b6e4345-38d3255d-8d8f-4b7f-927e-5a7330420b08" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 781.898378] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead29ec2-11a9-43a0-8f56-cce33b844a42 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.911241] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af36c67-8afb-496b-9955-c57d732cbaa6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.959843] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee5e898-6d18-4479-b776-165cbae9f369 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.971141] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b86f937-7cb9-4081-8e81-36d98d8e6b89 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.990994] env[65522]: DEBUG nova.compute.provider_tree [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.048882] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 782.069358] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 782.069553] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 782.081045] env[65522]: INFO nova.compute.manager [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Took 32.46 seconds to build instance. [ 782.256114] env[65522]: INFO nova.compute.manager [-] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Took 2.66 seconds to deallocate network for instance. [ 782.320305] env[65522]: DEBUG oslo_concurrency.lockutils [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.320305] env[65522]: DEBUG oslo_concurrency.lockutils [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 782.320678] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8593b69c-a6f2-472e-99a5-f45ed844f11b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.343262] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c0e3018-83fa-4b11-8644-8f8ffaffd650 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.369624] env[65522]: WARNING neutronclient.v2_0.client [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 782.380935] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Reconfiguring VM to detach interface {{(pid=65522) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 782.383539] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cdebf22d-3e1e-4f7f-9b1c-84d1722fe294 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.400947] env[65522]: DEBUG nova.network.neutron [-] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 782.410474] env[65522]: DEBUG oslo_vmware.api [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 782.410474] env[65522]: value = "task-5113961" [ 782.410474] env[65522]: _type = "Task" [ 782.410474] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.429193] env[65522]: DEBUG oslo_vmware.api [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113961, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.462764] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 782.494905] env[65522]: DEBUG nova.scheduler.client.report [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 782.583601] env[65522]: DEBUG oslo_concurrency.lockutils [None req-035424a3-7309-4a0a-bb21-c4865d2a5665 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "495fa98f-a9db-4214-87cc-d29209d3cb62" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.976s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 782.639500] env[65522]: DEBUG nova.compute.manager [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 782.676690] env[65522]: DEBUG nova.virt.hardware [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 782.677141] env[65522]: DEBUG nova.virt.hardware [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 782.677365] env[65522]: DEBUG nova.virt.hardware [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 782.677682] env[65522]: DEBUG nova.virt.hardware [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 782.677933] env[65522]: DEBUG nova.virt.hardware [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 782.678179] env[65522]: DEBUG nova.virt.hardware [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 782.679256] env[65522]: DEBUG nova.virt.hardware [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.679256] env[65522]: DEBUG nova.virt.hardware [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 782.679256] env[65522]: DEBUG nova.virt.hardware [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 782.679256] env[65522]: DEBUG nova.virt.hardware [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 782.680086] env[65522]: DEBUG nova.virt.hardware [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 782.680933] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717fd659-118a-4027-93fc-edc4a2065de8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.698468] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a4c8e1-3450-416e-b79b-aac186810e41 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.766097] env[65522]: DEBUG oslo_concurrency.lockutils [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 782.903166] env[65522]: INFO nova.compute.manager [-] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Took 2.39 seconds to deallocate network for instance. [ 782.918896] env[65522]: DEBUG nova.network.neutron [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Successfully created port: 012f75e3-8bea-4de0-9d56-d0b81de150d3 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 782.924856] env[65522]: DEBUG oslo_vmware.api [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113961, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.001874] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.386s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 783.004449] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 26.237s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.004777] env[65522]: DEBUG nova.objects.instance [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 783.036560] env[65522]: INFO nova.scheduler.client.report [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Deleted allocations for instance a09c12be-0aaa-4934-a1b3-5af29feba8a6 [ 783.301501] env[65522]: DEBUG nova.network.neutron [-] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 783.374689] env[65522]: WARNING neutronclient.v2_0.client [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.376607] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.377158] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.411401] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.424442] env[65522]: DEBUG oslo_vmware.api [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113961, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.546392] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbb54f1-df55-46a5-8788-8f183ac26612 tempest-ServersTestManualDisk-124804535 tempest-ServersTestManualDisk-124804535-project-member] Lock "a09c12be-0aaa-4934-a1b3-5af29feba8a6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.179s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 783.805356] env[65522]: INFO nova.compute.manager [-] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Took 2.55 seconds to deallocate network for instance. [ 783.870696] env[65522]: DEBUG nova.network.neutron [-] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 783.923248] env[65522]: DEBUG oslo_vmware.api [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113961, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.014921] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61ee3b9e-c3ba-43cf-8c79-3af1eaf127fd tempest-ServersAdmin275Test-1413200578 tempest-ServersAdmin275Test-1413200578-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 784.016125] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.452s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 784.016376] env[65522]: DEBUG nova.objects.instance [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lazy-loading 'resources' on Instance uuid 428873b8-1e58-4924-86f3-8feca8f53342 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 784.313112] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 784.373816] env[65522]: INFO nova.compute.manager [-] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Took 2.58 seconds to deallocate network for instance. [ 784.423616] env[65522]: DEBUG oslo_vmware.api [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113961, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.835207] env[65522]: DEBUG nova.network.neutron [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Updated VIF entry in instance network info cache for port 3d5e8380-9410-49a2-873c-962835dabfaa. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 784.836154] env[65522]: DEBUG nova.network.neutron [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Updating instance_info_cache with network_info: [{"id": "3d5e8380-9410-49a2-873c-962835dabfaa", "address": "fa:16:3e:1f:ae:89", "network": {"id": "0ab23355-59a7-409b-89f3-089fbd03db5e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-960579569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acbde84c090e4c0c919728b809c73d02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d5e8380-94", "ovs_interfaceid": "3d5e8380-9410-49a2-873c-962835dabfaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 784.886017] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 784.925566] env[65522]: DEBUG oslo_vmware.api [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113961, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.049335] env[65522]: DEBUG nova.network.neutron [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Successfully updated port: 012f75e3-8bea-4de0-9d56-d0b81de150d3 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 785.205021] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fc8f10-0629-4bd0-9583-249772dd35ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.211337] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8ff10b-e87a-496f-baf3-3c484fcb6200 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.256107] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c736edc-edc3-4f19-a32a-08ffc69f0a77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.265637] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dece8a5-d435-467f-b267-687aee58aa4f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.286159] env[65522]: DEBUG nova.compute.provider_tree [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.343570] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Releasing lock "refresh_cache-1e8088ef-6354-40a0-95ee-06732e7f088f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 785.345265] env[65522]: DEBUG nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Received event network-vif-plugged-f33077f4-9852-477a-ae48-9e86ad80fdce {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 785.345466] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Acquiring lock "24f520ec-6a56-4f17-9ae4-2c856bf99582-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 785.345779] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Lock "24f520ec-6a56-4f17-9ae4-2c856bf99582-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.346050] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Lock "24f520ec-6a56-4f17-9ae4-2c856bf99582-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 785.346245] env[65522]: DEBUG nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] No waiting events found dispatching network-vif-plugged-f33077f4-9852-477a-ae48-9e86ad80fdce {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 785.346416] env[65522]: WARNING nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Received unexpected event network-vif-plugged-f33077f4-9852-477a-ae48-9e86ad80fdce for instance with vm_state building and task_state spawning. [ 785.346585] env[65522]: DEBUG nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Received event network-changed-f33077f4-9852-477a-ae48-9e86ad80fdce {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 785.346739] env[65522]: DEBUG nova.compute.manager [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Refreshing instance network info cache due to event network-changed-f33077f4-9852-477a-ae48-9e86ad80fdce. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 785.346946] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Acquiring lock "refresh_cache-24f520ec-6a56-4f17-9ae4-2c856bf99582" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.347304] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Acquired lock "refresh_cache-24f520ec-6a56-4f17-9ae4-2c856bf99582" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 785.347402] env[65522]: DEBUG nova.network.neutron [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Refreshing network info cache for port f33077f4-9852-477a-ae48-9e86ad80fdce {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 785.425742] env[65522]: DEBUG oslo_vmware.api [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113961, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.508193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "1e8088ef-6354-40a0-95ee-06732e7f088f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 785.508193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "1e8088ef-6354-40a0-95ee-06732e7f088f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.508193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "1e8088ef-6354-40a0-95ee-06732e7f088f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 785.508193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "1e8088ef-6354-40a0-95ee-06732e7f088f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.508193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "1e8088ef-6354-40a0-95ee-06732e7f088f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 785.510504] env[65522]: INFO nova.compute.manager [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Terminating instance [ 785.553721] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "refresh_cache-7693f017-69b3-4ec4-ae3f-7e65ed34a471" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.553721] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquired lock "refresh_cache-7693f017-69b3-4ec4-ae3f-7e65ed34a471" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 785.553721] env[65522]: DEBUG nova.network.neutron [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 785.775031] env[65522]: DEBUG nova.compute.manager [req-1f683f27-6f41-415f-8f5b-31b198db864b req-09921b8a-bb0c-496b-ae67-9c8a2c2a6568 service nova] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Received event network-vif-deleted-97dc7703-dd4a-43dd-8e2d-02f8a31d1a69 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 785.791269] env[65522]: DEBUG nova.scheduler.client.report [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 785.853546] env[65522]: WARNING neutronclient.v2_0.client [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 785.854337] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.856273] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.927943] env[65522]: DEBUG oslo_vmware.api [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113961, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.020020] env[65522]: DEBUG nova.compute.manager [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 786.020020] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 786.020020] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dfb6e42-15f1-4c89-8dfc-ab351537aaaf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.029512] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 786.029512] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-abb471f7-4d7c-4d37-85cb-baccb493dee4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.040171] env[65522]: DEBUG oslo_vmware.api [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 786.040171] env[65522]: value = "task-5113962" [ 786.040171] env[65522]: _type = "Task" [ 786.040171] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.050514] env[65522]: DEBUG oslo_vmware.api [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113962, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.061847] env[65522]: WARNING openstack [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.061847] env[65522]: WARNING openstack [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.207850] env[65522]: DEBUG nova.network.neutron [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 786.226980] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.229712] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.297773] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.281s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 786.303849] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 24.337s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 786.304448] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 786.304827] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 786.305361] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.124s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 786.309930] env[65522]: INFO nova.compute.claims [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.317549] env[65522]: DEBUG nova.compute.manager [req-62977527-f388-4d8a-a74d-fced1a428a75 req-9e554295-d5da-44f8-bedb-bff570b56a7a service nova] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Received event network-vif-deleted-03d0a08a-a560-4f8a-8cc9-24211dff931b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 786.321169] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2e6c2f0-4c1c-4f74-98f3-ef4e09665d38 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.328942] env[65522]: WARNING openstack [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.329983] env[65522]: WARNING openstack [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.361396] env[65522]: INFO nova.scheduler.client.report [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Deleted allocations for instance 428873b8-1e58-4924-86f3-8feca8f53342 [ 786.365071] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c25088-dc2a-44b9-9514-d114fd7fd604 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.399757] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62dec340-4dc8-4834-be22-e9aed87be342 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.413023] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87c9b79-7dab-4611-85de-5c48abf6bb4d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.434067] env[65522]: DEBUG oslo_vmware.api [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113961, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.460354] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=177772MB free_disk=94GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 786.460590] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 786.555432] env[65522]: DEBUG oslo_vmware.api [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113962, 'name': PowerOffVM_Task, 'duration_secs': 0.235131} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.555751] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 786.555930] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 786.556224] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3bd1680f-7c31-447b-b740-352e71316eda {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.570996] env[65522]: WARNING neutronclient.v2_0.client [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 786.572260] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.572795] env[65522]: WARNING openstack [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.630653] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 786.630908] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 786.631133] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Deleting the datastore file [datastore2] 1e8088ef-6354-40a0-95ee-06732e7f088f {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 786.631443] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fa81687d-e746-4fdb-a409-2987d15f6542 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.639339] env[65522]: DEBUG oslo_vmware.api [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 786.639339] env[65522]: value = "task-5113964" [ 786.639339] env[65522]: _type = "Task" [ 786.639339] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.650829] env[65522]: DEBUG oslo_vmware.api [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113964, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.653328] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a96318-8dd3-14da-0ac1-e58f23083277/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 786.654285] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39735202-4af4-47b2-b2fa-1d959d08ef00 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.657853] env[65522]: WARNING neutronclient.v2_0.client [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 786.658508] env[65522]: WARNING openstack [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.658875] env[65522]: WARNING openstack [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.670897] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a96318-8dd3-14da-0ac1-e58f23083277/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 786.671080] env[65522]: ERROR oslo_vmware.rw_handles [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a96318-8dd3-14da-0ac1-e58f23083277/disk-0.vmdk due to incomplete transfer. [ 786.671337] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e3751b21-87f6-4468-9658-db113fe4617f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.680377] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a96318-8dd3-14da-0ac1-e58f23083277/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 786.680575] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Uploaded image 2b17de9f-63a4-459f-8af1-578cd934c183 to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 786.682821] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 786.683129] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-66a6df6d-077a-4e3e-8a31-fb06d7eeaf59 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.695169] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 786.695169] env[65522]: value = "task-5113965" [ 786.695169] env[65522]: _type = "Task" [ 786.695169] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.705403] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113965, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.884892] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17f66147-0641-445b-901a-9f959814964c tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "428873b8-1e58-4924-86f3-8feca8f53342" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.164s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 786.910763] env[65522]: DEBUG nova.network.neutron [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Updated VIF entry in instance network info cache for port f33077f4-9852-477a-ae48-9e86ad80fdce. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 786.911161] env[65522]: DEBUG nova.network.neutron [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Updating instance_info_cache with network_info: [{"id": "f33077f4-9852-477a-ae48-9e86ad80fdce", "address": "fa:16:3e:43:06:9c", "network": {"id": "dc9d8462-3539-4cc5-b21a-e1136a8df8e6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1580876213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "b5c14a856d454048856539895f220c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf33077f4-98", "ovs_interfaceid": "f33077f4-9852-477a-ae48-9e86ad80fdce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 786.937563] env[65522]: DEBUG oslo_vmware.api [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113961, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.989713] env[65522]: DEBUG nova.network.neutron [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Updating instance_info_cache with network_info: [{"id": "012f75e3-8bea-4de0-9d56-d0b81de150d3", "address": "fa:16:3e:79:f5:4a", "network": {"id": "472285e2-f5f3-4616-92f7-4c0144d833d4", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-945513036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2bd38f300424db78568e0b8d9982f27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap012f75e3-8b", "ovs_interfaceid": "012f75e3-8bea-4de0-9d56-d0b81de150d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 787.155768] env[65522]: DEBUG oslo_vmware.api [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5113964, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162061} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.155768] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.155768] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 787.155954] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 787.156410] env[65522]: INFO nova.compute.manager [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 787.156605] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 787.156804] env[65522]: DEBUG nova.compute.manager [-] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 787.156912] env[65522]: DEBUG nova.network.neutron [-] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 787.158242] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.159228] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.159228] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.205181] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113965, 'name': Destroy_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.286616] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.415964] env[65522]: DEBUG oslo_concurrency.lockutils [req-b4d99988-cda3-48f1-8a79-0cd6722e9018 req-be66e173-6e8f-49d6-9174-02aa629a9709 service nova] Releasing lock "refresh_cache-24f520ec-6a56-4f17-9ae4-2c856bf99582" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 787.436507] env[65522]: DEBUG oslo_vmware.api [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113961, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.494099] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Releasing lock "refresh_cache-7693f017-69b3-4ec4-ae3f-7e65ed34a471" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 787.494099] env[65522]: DEBUG nova.compute.manager [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Instance network_info: |[{"id": "012f75e3-8bea-4de0-9d56-d0b81de150d3", "address": "fa:16:3e:79:f5:4a", "network": {"id": "472285e2-f5f3-4616-92f7-4c0144d833d4", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-945513036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2bd38f300424db78568e0b8d9982f27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap012f75e3-8b", "ovs_interfaceid": "012f75e3-8bea-4de0-9d56-d0b81de150d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 787.494099] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:f5:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '012f75e3-8bea-4de0-9d56-d0b81de150d3', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 787.503522] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 787.503522] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 787.503522] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dff729b2-de66-48c1-b7fa-98790dbee6ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.532580] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Acquiring lock "18568baa-83de-4b54-addf-dee66211c8d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.532811] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Lock "18568baa-83de-4b54-addf-dee66211c8d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.534110] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 787.534110] env[65522]: value = "task-5113966" [ 787.534110] env[65522]: _type = "Task" [ 787.534110] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.546376] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113966, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.710037] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113965, 'name': Destroy_Task, 'duration_secs': 0.564627} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.712237] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Destroyed the VM [ 787.712859] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 787.714011] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d864caeb-f7bf-4931-a8d6-764116f37375 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.726023] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 787.726023] env[65522]: value = "task-5113967" [ 787.726023] env[65522]: _type = "Task" [ 787.726023] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.740558] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113967, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.939917] env[65522]: DEBUG oslo_vmware.api [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113961, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.035445] env[65522]: DEBUG nova.compute.manager [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 788.041691] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357c5057-a33c-4269-9480-1684fc732735 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.058948] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcf4bcf9-4e66-4e32-b480-a051ce2ef409 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.062976] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113966, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.104930] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c97194c1-2852-4dac-af06-24bb0645d6db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.116307] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da85d880-2c3c-4453-98f5-114ae7c1f001 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.132343] env[65522]: DEBUG nova.compute.provider_tree [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.214163] env[65522]: DEBUG nova.network.neutron [-] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 788.246050] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113967, 'name': RemoveSnapshot_Task} progress is 12%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.435246] env[65522]: DEBUG oslo_vmware.api [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113961, 'name': ReconfigVM_Task, 'duration_secs': 5.832414} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.435551] env[65522]: DEBUG oslo_concurrency.lockutils [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 788.435820] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Reconfigured VM to detach interface {{(pid=65522) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 788.436383] env[65522]: WARNING neutronclient.v2_0.client [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 788.436775] env[65522]: WARNING neutronclient.v2_0.client [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 788.437470] env[65522]: WARNING openstack [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.437888] env[65522]: WARNING openstack [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.548504] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113966, 'name': CreateVM_Task, 'duration_secs': 0.536962} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.550803] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 788.551505] env[65522]: WARNING neutronclient.v2_0.client [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 788.551879] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.552051] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 788.552376] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 788.552637] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19e0f24a-5994-476d-8175-9ebc6f3bd050 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.557690] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 788.557690] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522579e3-931c-a65f-1ac6-c5cadead7a8c" [ 788.557690] env[65522]: _type = "Task" [ 788.557690] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.566422] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522579e3-931c-a65f-1ac6-c5cadead7a8c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.567508] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.636268] env[65522]: DEBUG nova.scheduler.client.report [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 788.719344] env[65522]: INFO nova.compute.manager [-] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Took 1.56 seconds to deallocate network for instance. [ 788.739631] env[65522]: DEBUG oslo_vmware.api [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113967, 'name': RemoveSnapshot_Task, 'duration_secs': 0.620816} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.740237] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 788.740903] env[65522]: INFO nova.compute.manager [None req-0343d552-586c-437a-998c-082d7c678c98 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Took 19.44 seconds to snapshot the instance on the hypervisor. [ 789.070777] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522579e3-931c-a65f-1ac6-c5cadead7a8c, 'name': SearchDatastore_Task, 'duration_secs': 0.010421} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.071110] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 789.071359] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 789.071814] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.071960] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 789.072182] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 789.072469] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb3f08ae-88c2-409a-b368-f7db3410afcb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.083584] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 789.083584] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 789.084177] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1adfefd-5bcc-4cad-9e60-f3d723c63bae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.090560] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 789.090560] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5260c68e-b73d-5318-7450-05fa74a66b05" [ 789.090560] env[65522]: _type = "Task" [ 789.090560] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.101346] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5260c68e-b73d-5318-7450-05fa74a66b05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.142576] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.837s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 789.143134] env[65522]: DEBUG nova.compute.manager [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 789.146571] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.943s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 789.148375] env[65522]: INFO nova.compute.claims [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 789.229458] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 789.606800] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5260c68e-b73d-5318-7450-05fa74a66b05, 'name': SearchDatastore_Task, 'duration_secs': 0.010485} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.608375] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31d884c6-77b9-4943-8fc6-b514bd52c02e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.616039] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 789.616039] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5228ed08-56d1-5234-29d9-21f21b1c532f" [ 789.616039] env[65522]: _type = "Task" [ 789.616039] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.627718] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5228ed08-56d1-5234-29d9-21f21b1c532f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.653595] env[65522]: DEBUG nova.compute.utils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 789.666820] env[65522]: DEBUG nova.compute.manager [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 789.666820] env[65522]: DEBUG nova.network.neutron [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 789.666820] env[65522]: WARNING neutronclient.v2_0.client [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.666820] env[65522]: WARNING neutronclient.v2_0.client [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.666820] env[65522]: WARNING openstack [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.667267] env[65522]: WARNING openstack [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.133024] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5228ed08-56d1-5234-29d9-21f21b1c532f, 'name': SearchDatastore_Task, 'duration_secs': 0.011022} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.133024] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 790.133024] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 7693f017-69b3-4ec4-ae3f-7e65ed34a471/7693f017-69b3-4ec4-ae3f-7e65ed34a471.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 790.133024] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bfb239d7-62c6-4e30-80ef-5e96b488cf49 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.141228] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 790.141228] env[65522]: value = "task-5113968" [ 790.141228] env[65522]: _type = "Task" [ 790.141228] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.150766] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113968, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.164026] env[65522]: DEBUG nova.compute.manager [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 790.274186] env[65522]: DEBUG nova.policy [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a5bf5840867a428cab2ad0eb59970e56', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04fdbd83ceb54e5fbbf8fe00aa66cb32', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 790.591512] env[65522]: DEBUG oslo_concurrency.lockutils [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.591735] env[65522]: DEBUG oslo_concurrency.lockutils [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 790.591888] env[65522]: DEBUG nova.network.neutron [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 790.658263] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113968, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.753368] env[65522]: DEBUG nova.network.neutron [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Successfully created port: ad290ac3-4edd-4dd7-b683-0f0283c2dcc4 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 790.782025] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f38e6f-2dee-4683-a09c-47c8b7a21d6a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.792224] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b767e0d-0f63-4cc0-a698-62a02bda9938 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.828103] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3cc6797-1c76-4e3b-9f95-c966b72abb0f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.837181] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361704b8-d6c7-4e97-ad1f-377a5d581077 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.855409] env[65522]: DEBUG nova.compute.provider_tree [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.095636] env[65522]: WARNING neutronclient.v2_0.client [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 791.096389] env[65522]: WARNING openstack [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 791.096746] env[65522]: WARNING openstack [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.163022] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113968, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533143} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.163022] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 7693f017-69b3-4ec4-ae3f-7e65ed34a471/7693f017-69b3-4ec4-ae3f-7e65ed34a471.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 791.163022] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 791.163022] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d926c026-e133-4040-ba79-616f9fc68544 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.168514] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 791.168514] env[65522]: value = "task-5113969" [ 791.168514] env[65522]: _type = "Task" [ 791.168514] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.180505] env[65522]: DEBUG nova.compute.manager [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 791.182785] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113969, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.216222] env[65522]: DEBUG nova.virt.hardware [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 791.216521] env[65522]: DEBUG nova.virt.hardware [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 791.216679] env[65522]: DEBUG nova.virt.hardware [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 791.216865] env[65522]: DEBUG nova.virt.hardware [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 791.217025] env[65522]: DEBUG nova.virt.hardware [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 791.217321] env[65522]: DEBUG nova.virt.hardware [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 791.217576] env[65522]: DEBUG nova.virt.hardware [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 791.217792] env[65522]: DEBUG nova.virt.hardware [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 791.218015] env[65522]: DEBUG nova.virt.hardware [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 791.218318] env[65522]: DEBUG nova.virt.hardware [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 791.218543] env[65522]: DEBUG nova.virt.hardware [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 791.220827] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a82a140-d8f6-4260-af45-1749d474cb15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.228979] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc4079c7-4266-4869-a253-395b8114777b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.358999] env[65522]: DEBUG nova.scheduler.client.report [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 791.678855] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113969, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07911} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.678935] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 791.684019] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963e8413-bfdc-440e-882d-a0f0991e038d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.709299] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] 7693f017-69b3-4ec4-ae3f-7e65ed34a471/7693f017-69b3-4ec4-ae3f-7e65ed34a471.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 791.709798] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-890247ae-fe1c-4dba-a2e5-cd2ff3694e63 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.732659] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 791.732659] env[65522]: value = "task-5113970" [ 791.732659] env[65522]: _type = "Task" [ 791.732659] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.740588] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113970, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.862043] env[65522]: DEBUG nova.compute.manager [req-71aafb21-67f6-4cbf-940b-11da2c21825c req-c1025819-5ad3-4336-a43d-99ed3a77dd51 service nova] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Received event network-vif-deleted-a7a9fee0-67af-4446-b838-0cd8922465b9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 791.862451] env[65522]: DEBUG nova.compute.manager [req-71aafb21-67f6-4cbf-940b-11da2c21825c req-c1025819-5ad3-4336-a43d-99ed3a77dd51 service nova] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Received event network-vif-deleted-3d5e8380-9410-49a2-873c-962835dabfaa {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 791.864999] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.719s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 791.865589] env[65522]: DEBUG nova.compute.manager [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 791.871021] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.381s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 791.872276] env[65522]: INFO nova.compute.claims [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 792.245444] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113970, 'name': ReconfigVM_Task, 'duration_secs': 0.333997} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.245444] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Reconfigured VM instance instance-0000002c to attach disk [datastore1] 7693f017-69b3-4ec4-ae3f-7e65ed34a471/7693f017-69b3-4ec4-ae3f-7e65ed34a471.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 792.245444] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-209f350f-f92e-4f1b-bef9-638580f670f4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.253018] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 792.253018] env[65522]: value = "task-5113971" [ 792.253018] env[65522]: _type = "Task" [ 792.253018] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.257725] env[65522]: WARNING openstack [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.258274] env[65522]: WARNING openstack [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.272838] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113971, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.370058] env[65522]: WARNING neutronclient.v2_0.client [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 792.370058] env[65522]: WARNING openstack [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.370058] env[65522]: WARNING openstack [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.380256] env[65522]: DEBUG nova.compute.utils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 792.381881] env[65522]: DEBUG nova.compute.manager [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 792.382087] env[65522]: DEBUG nova.network.neutron [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 792.382403] env[65522]: WARNING neutronclient.v2_0.client [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 792.382695] env[65522]: WARNING neutronclient.v2_0.client [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 792.383509] env[65522]: WARNING openstack [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.383936] env[65522]: WARNING openstack [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.510191] env[65522]: DEBUG nova.compute.manager [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Received event network-vif-deleted-cf5b96ab-5cf5-4c14-9e3e-778bd9bdf35c {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 792.510492] env[65522]: DEBUG nova.compute.manager [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Received event network-vif-plugged-012f75e3-8bea-4de0-9d56-d0b81de150d3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 792.510528] env[65522]: DEBUG oslo_concurrency.lockutils [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] Acquiring lock "7693f017-69b3-4ec4-ae3f-7e65ed34a471-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 792.510858] env[65522]: DEBUG oslo_concurrency.lockutils [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] Lock "7693f017-69b3-4ec4-ae3f-7e65ed34a471-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 792.510858] env[65522]: DEBUG oslo_concurrency.lockutils [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] Lock "7693f017-69b3-4ec4-ae3f-7e65ed34a471-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.511042] env[65522]: DEBUG nova.compute.manager [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] No waiting events found dispatching network-vif-plugged-012f75e3-8bea-4de0-9d56-d0b81de150d3 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 792.511189] env[65522]: WARNING nova.compute.manager [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Received unexpected event network-vif-plugged-012f75e3-8bea-4de0-9d56-d0b81de150d3 for instance with vm_state building and task_state spawning. [ 792.511341] env[65522]: DEBUG nova.compute.manager [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Received event network-changed-012f75e3-8bea-4de0-9d56-d0b81de150d3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 792.511481] env[65522]: DEBUG nova.compute.manager [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Refreshing instance network info cache due to event network-changed-012f75e3-8bea-4de0-9d56-d0b81de150d3. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 792.511792] env[65522]: DEBUG oslo_concurrency.lockutils [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] Acquiring lock "refresh_cache-7693f017-69b3-4ec4-ae3f-7e65ed34a471" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.511792] env[65522]: DEBUG oslo_concurrency.lockutils [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] Acquired lock "refresh_cache-7693f017-69b3-4ec4-ae3f-7e65ed34a471" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 792.512179] env[65522]: DEBUG nova.network.neutron [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Refreshing network info cache for port 012f75e3-8bea-4de0-9d56-d0b81de150d3 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 792.596960] env[65522]: DEBUG nova.policy [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '767535aa17a84adeb722a33713b7aa68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b8112ba49034e6fabcb3fbbd46edf41', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 792.642139] env[65522]: INFO nova.network.neutron [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Port 38d3255d-8d8f-4b7f-927e-5a7330420b08 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 792.642767] env[65522]: DEBUG nova.network.neutron [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Updating instance_info_cache with network_info: [{"id": "d24163e8-4f08-4908-9307-c4edb0d4ffb7", "address": "fa:16:3e:e2:90:31", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd24163e8-4f", "ovs_interfaceid": "d24163e8-4f08-4908-9307-c4edb0d4ffb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 792.764896] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113971, 'name': Rename_Task, 'duration_secs': 0.160077} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.765479] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 792.765747] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a969ae4-e3f4-4f84-a005-1d44f46ea908 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.770361] env[65522]: DEBUG nova.network.neutron [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Successfully updated port: ad290ac3-4edd-4dd7-b683-0f0283c2dcc4 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 792.780949] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 792.780949] env[65522]: value = "task-5113972" [ 792.780949] env[65522]: _type = "Task" [ 792.780949] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.801934] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113972, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.896337] env[65522]: DEBUG nova.compute.manager [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 793.015953] env[65522]: WARNING neutronclient.v2_0.client [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 793.017130] env[65522]: WARNING openstack [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.017480] env[65522]: WARNING openstack [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.052381] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "0b48b747-149c-4163-bdd0-7d0b07ddacfd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 793.052626] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "0b48b747-149c-4163-bdd0-7d0b07ddacfd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 793.146046] env[65522]: DEBUG oslo_concurrency.lockutils [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "refresh_cache-b2d70982-54e7-459c-a0d8-48bf4b6e4345" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 793.200753] env[65522]: DEBUG nova.network.neutron [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Successfully created port: e12c2307-4963-4a5a-af0f-19369217f430 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 793.277855] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Acquiring lock "refresh_cache-d2fdffb8-76d1-4b74-819f-25eb29ca91ce" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.278110] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Acquired lock "refresh_cache-d2fdffb8-76d1-4b74-819f-25eb29ca91ce" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 793.278271] env[65522]: DEBUG nova.network.neutron [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 793.297740] env[65522]: DEBUG oslo_vmware.api [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113972, 'name': PowerOnVM_Task, 'duration_secs': 0.491055} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.302026] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 793.302026] env[65522]: INFO nova.compute.manager [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Took 10.66 seconds to spawn the instance on the hypervisor. [ 793.302026] env[65522]: DEBUG nova.compute.manager [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 793.302026] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea7a46f-eb8b-4055-94a3-0e30756c8b63 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.344678] env[65522]: WARNING openstack [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.345465] env[65522]: WARNING openstack [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.544796] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb53118-e67a-45a0-9daa-bef10b0c6871 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.553203] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d053be-e1bf-485c-aaa7-37f8decc42b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.560060] env[65522]: DEBUG nova.compute.manager [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 793.591299] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c7e5537-28e1-4793-a85b-ffb9af8d61fb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.601439] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-237f4f8c-1f42-4dbe-8e8b-ee712748f0b3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.617944] env[65522]: DEBUG nova.compute.provider_tree [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.651242] env[65522]: DEBUG oslo_concurrency.lockutils [None req-64222b5a-4623-4749-9fd2-0278b2174199 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-b2d70982-54e7-459c-a0d8-48bf4b6e4345-38d3255d-8d8f-4b7f-927e-5a7330420b08" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 11.835s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 793.781734] env[65522]: WARNING openstack [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.782182] env[65522]: WARNING openstack [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.828880] env[65522]: INFO nova.compute.manager [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Took 40.83 seconds to build instance. [ 793.909309] env[65522]: DEBUG nova.compute.manager [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 793.945806] env[65522]: DEBUG nova.virt.hardware [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 793.946645] env[65522]: DEBUG nova.virt.hardware [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 793.946645] env[65522]: DEBUG nova.virt.hardware [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 793.946645] env[65522]: DEBUG nova.virt.hardware [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 793.946803] env[65522]: DEBUG nova.virt.hardware [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 793.946854] env[65522]: DEBUG nova.virt.hardware [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 793.947078] env[65522]: DEBUG nova.virt.hardware [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 793.947298] env[65522]: DEBUG nova.virt.hardware [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 793.947792] env[65522]: DEBUG nova.virt.hardware [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 793.947792] env[65522]: DEBUG nova.virt.hardware [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 793.947792] env[65522]: DEBUG nova.virt.hardware [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 793.948713] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f503d0b9-24dc-4800-acd0-6fa9894b7222 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.958724] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c98443-c38d-4a0e-8927-8d6d53097848 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.113425] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 794.121638] env[65522]: DEBUG nova.scheduler.client.report [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 794.265022] env[65522]: DEBUG nova.network.neutron [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 794.331421] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5b82e43-723b-434c-8cf2-5b9a989e4781 tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "7693f017-69b3-4ec4-ae3f-7e65ed34a471" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.344s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 794.628049] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.757s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 794.628670] env[65522]: DEBUG nova.compute.manager [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 794.631358] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.193s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 794.633530] env[65522]: INFO nova.compute.claims [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 794.843673] env[65522]: DEBUG nova.network.neutron [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Successfully updated port: e12c2307-4963-4a5a-af0f-19369217f430 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 794.937249] env[65522]: WARNING neutronclient.v2_0.client [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.937940] env[65522]: WARNING openstack [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.938473] env[65522]: WARNING openstack [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.986848] env[65522]: WARNING openstack [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.987265] env[65522]: WARNING openstack [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 795.138124] env[65522]: DEBUG nova.network.neutron [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Updated VIF entry in instance network info cache for port 012f75e3-8bea-4de0-9d56-d0b81de150d3. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 795.138521] env[65522]: DEBUG nova.network.neutron [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Updating instance_info_cache with network_info: [{"id": "012f75e3-8bea-4de0-9d56-d0b81de150d3", "address": "fa:16:3e:79:f5:4a", "network": {"id": "472285e2-f5f3-4616-92f7-4c0144d833d4", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-945513036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2bd38f300424db78568e0b8d9982f27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap012f75e3-8b", "ovs_interfaceid": "012f75e3-8bea-4de0-9d56-d0b81de150d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 795.141040] env[65522]: DEBUG nova.compute.utils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 795.145043] env[65522]: DEBUG nova.compute.manager [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 795.145292] env[65522]: DEBUG nova.network.neutron [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 795.145840] env[65522]: WARNING neutronclient.v2_0.client [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 795.146316] env[65522]: WARNING neutronclient.v2_0.client [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 795.146957] env[65522]: WARNING openstack [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 795.147392] env[65522]: WARNING openstack [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 795.156691] env[65522]: WARNING neutronclient.v2_0.client [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 795.159296] env[65522]: WARNING openstack [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 795.159296] env[65522]: WARNING openstack [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 795.215691] env[65522]: DEBUG nova.policy [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1dc7c40c5104394bb2c1293904246e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ab0a321a59248a286c9d8775c7a9ddc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 795.339313] env[65522]: DEBUG nova.network.neutron [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Updating instance_info_cache with network_info: [{"id": "ad290ac3-4edd-4dd7-b683-0f0283c2dcc4", "address": "fa:16:3e:e9:2b:6d", "network": {"id": "2e4ac13b-587a-4147-96fe-a0dbacab7ebd", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1030195272-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04fdbd83ceb54e5fbbf8fe00aa66cb32", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad290ac3-4e", "ovs_interfaceid": "ad290ac3-4edd-4dd7-b683-0f0283c2dcc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 795.348731] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.348731] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 795.348731] env[65522]: DEBUG nova.network.neutron [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 795.569589] env[65522]: DEBUG nova.network.neutron [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Successfully created port: 7e072218-2d88-437a-b6fb-597ced883532 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 795.646484] env[65522]: DEBUG oslo_concurrency.lockutils [req-2eb71641-a067-4f52-a4b8-11808cfbe64f req-18cfdc58-34a6-4496-89cf-e65b8516c18f service nova] Releasing lock "refresh_cache-7693f017-69b3-4ec4-ae3f-7e65ed34a471" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 795.656727] env[65522]: DEBUG nova.compute.manager [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 795.842838] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Releasing lock "refresh_cache-d2fdffb8-76d1-4b74-819f-25eb29ca91ce" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 795.843116] env[65522]: DEBUG nova.compute.manager [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Instance network_info: |[{"id": "ad290ac3-4edd-4dd7-b683-0f0283c2dcc4", "address": "fa:16:3e:e9:2b:6d", "network": {"id": "2e4ac13b-587a-4147-96fe-a0dbacab7ebd", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1030195272-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04fdbd83ceb54e5fbbf8fe00aa66cb32", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad290ac3-4e", "ovs_interfaceid": "ad290ac3-4edd-4dd7-b683-0f0283c2dcc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 795.843986] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e9:2b:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ad290ac3-4edd-4dd7-b683-0f0283c2dcc4', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 795.852145] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Creating folder: Project (04fdbd83ceb54e5fbbf8fe00aa66cb32). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 795.853482] env[65522]: WARNING openstack [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 795.853875] env[65522]: WARNING openstack [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 795.862126] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bdd3579e-b185-4822-99ee-03dde3f2d5e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.880432] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Created folder: Project (04fdbd83ceb54e5fbbf8fe00aa66cb32) in parent group-v994660. [ 795.880709] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Creating folder: Instances. Parent ref: group-v994780. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 795.885226] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0345f9b0-bf22-427e-9ba1-f086976cb52a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.898334] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Created folder: Instances in parent group-v994780. [ 795.898880] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 795.899514] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 795.899911] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1dea42fb-9d82-44d9-b771-81793b269c53 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.929990] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 795.929990] env[65522]: value = "task-5113975" [ 795.929990] env[65522]: _type = "Task" [ 795.929990] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.945158] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113975, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.953257] env[65522]: DEBUG nova.network.neutron [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 796.020992] env[65522]: WARNING openstack [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 796.021390] env[65522]: WARNING openstack [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 796.125530] env[65522]: WARNING neutronclient.v2_0.client [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 796.126198] env[65522]: WARNING openstack [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 796.126556] env[65522]: WARNING openstack [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 796.264809] env[65522]: DEBUG nova.network.neutron [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Updating instance_info_cache with network_info: [{"id": "e12c2307-4963-4a5a-af0f-19369217f430", "address": "fa:16:3e:9f:b8:7d", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape12c2307-49", "ovs_interfaceid": "e12c2307-4963-4a5a-af0f-19369217f430", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 796.353258] env[65522]: DEBUG nova.compute.manager [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Received event network-vif-deleted-38d3255d-8d8f-4b7f-927e-5a7330420b08 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 796.353489] env[65522]: DEBUG nova.compute.manager [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Received event network-vif-plugged-ad290ac3-4edd-4dd7-b683-0f0283c2dcc4 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 796.353850] env[65522]: DEBUG oslo_concurrency.lockutils [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] Acquiring lock "d2fdffb8-76d1-4b74-819f-25eb29ca91ce-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 796.353850] env[65522]: DEBUG oslo_concurrency.lockutils [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] Lock "d2fdffb8-76d1-4b74-819f-25eb29ca91ce-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 796.354071] env[65522]: DEBUG oslo_concurrency.lockutils [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] Lock "d2fdffb8-76d1-4b74-819f-25eb29ca91ce-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.354254] env[65522]: DEBUG nova.compute.manager [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] No waiting events found dispatching network-vif-plugged-ad290ac3-4edd-4dd7-b683-0f0283c2dcc4 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 796.354457] env[65522]: WARNING nova.compute.manager [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Received unexpected event network-vif-plugged-ad290ac3-4edd-4dd7-b683-0f0283c2dcc4 for instance with vm_state building and task_state spawning. [ 796.354575] env[65522]: DEBUG nova.compute.manager [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Received event network-changed-ad290ac3-4edd-4dd7-b683-0f0283c2dcc4 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 796.354727] env[65522]: DEBUG nova.compute.manager [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Refreshing instance network info cache due to event network-changed-ad290ac3-4edd-4dd7-b683-0f0283c2dcc4. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 796.355116] env[65522]: DEBUG oslo_concurrency.lockutils [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] Acquiring lock "refresh_cache-d2fdffb8-76d1-4b74-819f-25eb29ca91ce" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.355116] env[65522]: DEBUG oslo_concurrency.lockutils [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] Acquired lock "refresh_cache-d2fdffb8-76d1-4b74-819f-25eb29ca91ce" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 796.355233] env[65522]: DEBUG nova.network.neutron [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Refreshing network info cache for port ad290ac3-4edd-4dd7-b683-0f0283c2dcc4 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 796.418509] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6e5cb5-aa88-41b1-b713-2eaeec21289d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.436568] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6168f5-a5cc-49e5-b290-6e2f3ff069c5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.481833] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113975, 'name': CreateVM_Task, 'duration_secs': 0.3891} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.482178] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 796.483579] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-244196d8-8b79-43d1-982e-d3f761ef12fd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.486301] env[65522]: WARNING neutronclient.v2_0.client [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 796.486652] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.486803] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 796.487178] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 796.487467] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa81614a-cbf7-4bbd-99db-9cc505b73727 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.497163] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c2e2e4a-c349-4849-a53f-21172929d030 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.501856] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Waiting for the task: (returnval){ [ 796.501856] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52780a8f-e669-092b-5d00-f74ef0d26264" [ 796.501856] env[65522]: _type = "Task" [ 796.501856] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.516460] env[65522]: DEBUG nova.compute.provider_tree [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.524626] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52780a8f-e669-092b-5d00-f74ef0d26264, 'name': SearchDatastore_Task, 'duration_secs': 0.025273} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.524626] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 796.524626] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 796.524626] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.524941] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 796.524941] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 796.525347] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7dc0732-bf5f-4a00-a085-799e5bc2dc1a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.539156] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 796.539156] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 796.539896] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44d7b9a8-2bff-44f3-9f1c-1a14f2872b7e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.549516] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Waiting for the task: (returnval){ [ 796.549516] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526c9bc2-4a66-43cf-527d-5fe7ea168f2a" [ 796.549516] env[65522]: _type = "Task" [ 796.549516] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.557161] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526c9bc2-4a66-43cf-527d-5fe7ea168f2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.597612] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Acquiring lock "880390f5-00f1-4dea-9345-29c06289b340" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 796.597612] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Lock "880390f5-00f1-4dea-9345-29c06289b340" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 796.667470] env[65522]: DEBUG nova.compute.manager [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 796.700683] env[65522]: DEBUG nova.virt.hardware [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 796.700683] env[65522]: DEBUG nova.virt.hardware [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 796.700683] env[65522]: DEBUG nova.virt.hardware [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 796.700832] env[65522]: DEBUG nova.virt.hardware [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 796.700913] env[65522]: DEBUG nova.virt.hardware [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 796.701950] env[65522]: DEBUG nova.virt.hardware [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 796.701950] env[65522]: DEBUG nova.virt.hardware [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 796.701950] env[65522]: DEBUG nova.virt.hardware [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 796.701950] env[65522]: DEBUG nova.virt.hardware [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 796.701950] env[65522]: DEBUG nova.virt.hardware [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 796.701950] env[65522]: DEBUG nova.virt.hardware [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 796.702816] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a73b95a-aa32-4a64-803c-0c9ca361db27 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.711828] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c1f294-183a-4115-8a7d-11b62e99f6b4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.771544] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 796.771911] env[65522]: DEBUG nova.compute.manager [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Instance network_info: |[{"id": "e12c2307-4963-4a5a-af0f-19369217f430", "address": "fa:16:3e:9f:b8:7d", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape12c2307-49", "ovs_interfaceid": "e12c2307-4963-4a5a-af0f-19369217f430", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 796.772410] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:b8:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4b033f4d-2e92-4702-add6-410a29d3f251', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e12c2307-4963-4a5a-af0f-19369217f430', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 796.782109] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 796.782375] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 796.782619] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed18e609-7934-4955-83c2-40b2df9a2752 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.804870] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 796.804870] env[65522]: value = "task-5113976" [ 796.804870] env[65522]: _type = "Task" [ 796.804870] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.815134] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113976, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.857702] env[65522]: WARNING neutronclient.v2_0.client [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 796.858486] env[65522]: WARNING openstack [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 796.858853] env[65522]: WARNING openstack [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 797.019323] env[65522]: DEBUG nova.scheduler.client.report [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 797.059520] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526c9bc2-4a66-43cf-527d-5fe7ea168f2a, 'name': SearchDatastore_Task, 'duration_secs': 0.020444} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.060602] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fdb357aa-1a5e-4493-bccd-c6eacb1d7114 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.067623] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Waiting for the task: (returnval){ [ 797.067623] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5268369a-dddf-732d-3d12-8ef8bc20f39b" [ 797.067623] env[65522]: _type = "Task" [ 797.067623] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.078835] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5268369a-dddf-732d-3d12-8ef8bc20f39b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.100203] env[65522]: DEBUG nova.compute.manager [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 797.191820] env[65522]: DEBUG nova.network.neutron [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Successfully updated port: 7e072218-2d88-437a-b6fb-597ced883532 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 797.316743] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113976, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.441210] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 797.441649] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 797.441836] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 797.442170] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 797.442240] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 797.445422] env[65522]: INFO nova.compute.manager [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Terminating instance [ 797.533589] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.902s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 797.534845] env[65522]: DEBUG nova.compute.manager [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 797.537850] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 21.994s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 797.583405] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5268369a-dddf-732d-3d12-8ef8bc20f39b, 'name': SearchDatastore_Task, 'duration_secs': 0.014973} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.583405] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 797.583405] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] d2fdffb8-76d1-4b74-819f-25eb29ca91ce/d2fdffb8-76d1-4b74-819f-25eb29ca91ce.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 797.583405] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7dad5e75-19a9-4643-b6d2-8d340697721a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.594299] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Waiting for the task: (returnval){ [ 797.594299] env[65522]: value = "task-5113977" [ 797.594299] env[65522]: _type = "Task" [ 797.594299] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.608661] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5113977, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.632488] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 797.643917] env[65522]: WARNING openstack [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 797.644322] env[65522]: WARNING openstack [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 797.700844] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquiring lock "refresh_cache-9fa6e208-f913-4ecc-8548-ed6503fe6273" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.701078] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquired lock "refresh_cache-9fa6e208-f913-4ecc-8548-ed6503fe6273" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 797.701264] env[65522]: DEBUG nova.network.neutron [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 797.740061] env[65522]: WARNING neutronclient.v2_0.client [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 797.740896] env[65522]: WARNING openstack [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 797.741202] env[65522]: WARNING openstack [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 797.819638] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113976, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.866550] env[65522]: DEBUG nova.network.neutron [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Updated VIF entry in instance network info cache for port ad290ac3-4edd-4dd7-b683-0f0283c2dcc4. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 797.866550] env[65522]: DEBUG nova.network.neutron [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Updating instance_info_cache with network_info: [{"id": "ad290ac3-4edd-4dd7-b683-0f0283c2dcc4", "address": "fa:16:3e:e9:2b:6d", "network": {"id": "2e4ac13b-587a-4147-96fe-a0dbacab7ebd", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1030195272-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04fdbd83ceb54e5fbbf8fe00aa66cb32", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad290ac3-4e", "ovs_interfaceid": "ad290ac3-4edd-4dd7-b683-0f0283c2dcc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 797.955038] env[65522]: DEBUG nova.compute.manager [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 797.955329] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 797.959910] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e869c777-0848-411d-84fb-8c80f4b7069f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.971147] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 797.971147] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ea2890e-93be-45a2-9cd3-548fbdf572cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.982184] env[65522]: DEBUG oslo_vmware.api [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 797.982184] env[65522]: value = "task-5113978" [ 797.982184] env[65522]: _type = "Task" [ 797.982184] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.996630] env[65522]: DEBUG oslo_vmware.api [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113978, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.047032] env[65522]: DEBUG nova.compute.utils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 798.056447] env[65522]: DEBUG nova.compute.manager [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 798.056447] env[65522]: DEBUG nova.network.neutron [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 798.056447] env[65522]: WARNING neutronclient.v2_0.client [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.056946] env[65522]: WARNING neutronclient.v2_0.client [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.057713] env[65522]: WARNING openstack [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.060926] env[65522]: WARNING openstack [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.115124] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5113977, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.129590] env[65522]: DEBUG nova.policy [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc0c40934d3848b78cad1f3da2245cf0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ecf042e8e8245cdb7e58d41b80e6737', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 798.207945] env[65522]: WARNING openstack [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.208474] env[65522]: WARNING openstack [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.281841] env[65522]: DEBUG nova.network.neutron [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 798.313059] env[65522]: WARNING openstack [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.313453] env[65522]: WARNING openstack [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.327521] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113976, 'name': CreateVM_Task, 'duration_secs': 1.514717} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.327696] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 798.328200] env[65522]: WARNING neutronclient.v2_0.client [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.328553] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.328706] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 798.329020] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 798.329294] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bbc999c-c96d-4ae9-b1af-8000c43677af {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.334541] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 798.334541] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e267e6-0604-8d79-ca5a-3ff61e6b1fd2" [ 798.334541] env[65522]: _type = "Task" [ 798.334541] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.346747] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e267e6-0604-8d79-ca5a-3ff61e6b1fd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.369366] env[65522]: DEBUG oslo_concurrency.lockutils [req-61f29876-3a59-4266-b0e3-37eb4bee5143 req-968fa24e-3e64-4e82-b444-dcec85a38933 service nova] Releasing lock "refresh_cache-d2fdffb8-76d1-4b74-819f-25eb29ca91ce" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 798.436069] env[65522]: WARNING neutronclient.v2_0.client [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.437354] env[65522]: WARNING openstack [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.437354] env[65522]: WARNING openstack [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.498039] env[65522]: DEBUG oslo_vmware.api [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113978, 'name': PowerOffVM_Task, 'duration_secs': 0.288552} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.499546] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 798.499657] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 798.499953] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2dac3e4-186e-4553-96ca-bd8d3932848d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.503053] env[65522]: DEBUG nova.network.neutron [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Successfully created port: 4c16f1be-de7b-45f4-9510-b1e2d358c37c {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 798.545243] env[65522]: DEBUG nova.network.neutron [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Updating instance_info_cache with network_info: [{"id": "7e072218-2d88-437a-b6fb-597ced883532", "address": "fa:16:3e:21:b1:bf", "network": {"id": "0ab7f06c-0131-43cd-8161-2b87ea9cfb91", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-160966193-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "6ab0a321a59248a286c9d8775c7a9ddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e072218-2d", "ovs_interfaceid": "7e072218-2d88-437a-b6fb-597ced883532", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 798.556164] env[65522]: DEBUG nova.compute.manager [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 798.582196] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 798.582499] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 798.582688] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Deleting the datastore file [datastore1] b2d70982-54e7-459c-a0d8-48bf4b6e4345 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 798.583107] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df621066-73ae-40d9-90a6-fba754a045c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.595026] env[65522]: DEBUG oslo_vmware.api [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 798.595026] env[65522]: value = "task-5113980" [ 798.595026] env[65522]: _type = "Task" [ 798.595026] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.617339] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5113977, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.773879} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.618738] env[65522]: DEBUG oslo_vmware.api [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113980, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.618738] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] d2fdffb8-76d1-4b74-819f-25eb29ca91ce/d2fdffb8-76d1-4b74-819f-25eb29ca91ce.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 798.618738] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 798.618879] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9bf34544-6663-48c8-95ae-10b77213d758 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.633553] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Waiting for the task: (returnval){ [ 798.633553] env[65522]: value = "task-5113981" [ 798.633553] env[65522]: _type = "Task" [ 798.633553] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.645808] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5113981, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.730520] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 798.731100] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.819542] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0dc4d64-533d-4b90-a786-9d99d8e625e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.829429] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba959f2e-047e-4a9f-aa5d-847d63d8bde1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.868318] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a67fd1-1215-4abe-ab9f-ad8677671cb8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.876560] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e267e6-0604-8d79-ca5a-3ff61e6b1fd2, 'name': SearchDatastore_Task, 'duration_secs': 0.030432} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.877265] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 798.877517] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 798.877779] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.878041] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 798.878255] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 798.878586] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b1254de-f702-4835-8efb-608395649a7b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.885689] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03dbd409-db90-4525-bd92-e7b56fe09b2d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.894248] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 798.894437] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 798.903126] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72bee4ae-8fc9-48cd-ae80-c8aaaca05c8e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.906260] env[65522]: DEBUG nova.compute.provider_tree [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.915371] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 798.915371] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520e4f6d-e3af-d97b-0401-fe361ed72d87" [ 798.915371] env[65522]: _type = "Task" [ 798.915371] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.924196] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520e4f6d-e3af-d97b-0401-fe361ed72d87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.050582] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Releasing lock "refresh_cache-9fa6e208-f913-4ecc-8548-ed6503fe6273" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 799.050850] env[65522]: DEBUG nova.compute.manager [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Instance network_info: |[{"id": "7e072218-2d88-437a-b6fb-597ced883532", "address": "fa:16:3e:21:b1:bf", "network": {"id": "0ab7f06c-0131-43cd-8161-2b87ea9cfb91", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-160966193-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "6ab0a321a59248a286c9d8775c7a9ddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e072218-2d", "ovs_interfaceid": "7e072218-2d88-437a-b6fb-597ced883532", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 799.051418] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:b1:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1eed7865-f9d8-463e-843f-3b0b3a962a2c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e072218-2d88-437a-b6fb-597ced883532', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 799.059643] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Creating folder: Project (6ab0a321a59248a286c9d8775c7a9ddc). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 799.064394] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1e42bb0b-f9ac-4b85-b803-bab167145bf5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.077600] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Created folder: Project (6ab0a321a59248a286c9d8775c7a9ddc) in parent group-v994660. [ 799.077791] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Creating folder: Instances. Parent ref: group-v994784. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 799.078068] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8f0a9293-9e79-444e-8a6d-60dd206a4ee0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.088445] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Created folder: Instances in parent group-v994784. [ 799.088748] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 799.088919] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 799.089179] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8c92ab5-2a6a-449a-8937-842c5797e72e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.116992] env[65522]: DEBUG oslo_vmware.api [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113980, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.118314] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 799.118314] env[65522]: value = "task-5113984" [ 799.118314] env[65522]: _type = "Task" [ 799.118314] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.145589] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5113981, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080153} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.145589] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 799.145589] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3289fd59-1ec7-4dd7-895d-88363dd7f074 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.170937] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] d2fdffb8-76d1-4b74-819f-25eb29ca91ce/d2fdffb8-76d1-4b74-819f-25eb29ca91ce.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 799.171341] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e97998f5-30e3-44fd-9d86-1f20fac25560 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.193063] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Waiting for the task: (returnval){ [ 799.193063] env[65522]: value = "task-5113985" [ 799.193063] env[65522]: _type = "Task" [ 799.193063] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.203800] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5113985, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.234720] env[65522]: DEBUG nova.compute.manager [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 799.411656] env[65522]: DEBUG nova.scheduler.client.report [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 799.426348] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520e4f6d-e3af-d97b-0401-fe361ed72d87, 'name': SearchDatastore_Task, 'duration_secs': 0.039547} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.427458] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23192909-1fd7-4b92-94f0-8c4a7513ae77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.434688] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 799.434688] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52820189-bf4d-f026-a13b-d66179b9a36e" [ 799.434688] env[65522]: _type = "Task" [ 799.434688] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.448824] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52820189-bf4d-f026-a13b-d66179b9a36e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.569059] env[65522]: DEBUG nova.compute.manager [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 799.593379] env[65522]: DEBUG nova.virt.hardware [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 799.593644] env[65522]: DEBUG nova.virt.hardware [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 799.593800] env[65522]: DEBUG nova.virt.hardware [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 799.593978] env[65522]: DEBUG nova.virt.hardware [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 799.594138] env[65522]: DEBUG nova.virt.hardware [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 799.594285] env[65522]: DEBUG nova.virt.hardware [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 799.594499] env[65522]: DEBUG nova.virt.hardware [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 799.594655] env[65522]: DEBUG nova.virt.hardware [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 799.594819] env[65522]: DEBUG nova.virt.hardware [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 799.594976] env[65522]: DEBUG nova.virt.hardware [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 799.595165] env[65522]: DEBUG nova.virt.hardware [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 799.596205] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94a1a7c-2b49-4d96-812e-86b8624d1d85 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.605789] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf614269-48d9-4366-bf17-9690aac6370a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.628207] env[65522]: DEBUG oslo_vmware.api [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5113980, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.706388} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.632275] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 799.632275] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 799.632416] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 799.632517] env[65522]: INFO nova.compute.manager [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Took 1.68 seconds to destroy the instance on the hypervisor. [ 799.633901] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 799.633901] env[65522]: DEBUG nova.compute.manager [-] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 799.633901] env[65522]: DEBUG nova.network.neutron [-] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 799.633901] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 799.633901] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 799.634129] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 799.644035] env[65522]: DEBUG nova.compute.manager [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Received event network-vif-plugged-e12c2307-4963-4a5a-af0f-19369217f430 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 799.644290] env[65522]: DEBUG oslo_concurrency.lockutils [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Acquiring lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.644479] env[65522]: DEBUG oslo_concurrency.lockutils [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 799.644636] env[65522]: DEBUG oslo_concurrency.lockutils [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 799.644796] env[65522]: DEBUG nova.compute.manager [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] No waiting events found dispatching network-vif-plugged-e12c2307-4963-4a5a-af0f-19369217f430 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 799.644950] env[65522]: WARNING nova.compute.manager [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Received unexpected event network-vif-plugged-e12c2307-4963-4a5a-af0f-19369217f430 for instance with vm_state building and task_state spawning. [ 799.645124] env[65522]: DEBUG nova.compute.manager [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Received event network-changed-e12c2307-4963-4a5a-af0f-19369217f430 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 799.645269] env[65522]: DEBUG nova.compute.manager [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Refreshing instance network info cache due to event network-changed-e12c2307-4963-4a5a-af0f-19369217f430. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 799.645440] env[65522]: DEBUG oslo_concurrency.lockutils [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Acquiring lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.645573] env[65522]: DEBUG oslo_concurrency.lockutils [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Acquired lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 799.645808] env[65522]: DEBUG nova.network.neutron [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Refreshing network info cache for port e12c2307-4963-4a5a-af0f-19369217f430 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 799.655189] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113984, 'name': CreateVM_Task, 'duration_secs': 0.397789} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.655189] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 799.655189] env[65522]: WARNING neutronclient.v2_0.client [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 799.655843] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.655843] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 799.655843] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 799.659031] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c309d9eb-c92b-425e-b951-c9df9c0973d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.664842] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Waiting for the task: (returnval){ [ 799.664842] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521c87dc-8f6e-8abe-8cdf-e8cced9fca18" [ 799.664842] env[65522]: _type = "Task" [ 799.664842] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.677809] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521c87dc-8f6e-8abe-8cdf-e8cced9fca18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.705825] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5113985, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.721384] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 799.755862] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.769411] env[65522]: DEBUG nova.compute.manager [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 799.771329] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7bffaf2-cae7-47f0-9de3-8f36d5872f4b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.947723] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52820189-bf4d-f026-a13b-d66179b9a36e, 'name': SearchDatastore_Task, 'duration_secs': 0.023319} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.948032] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 799.948434] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 1eebb320-30e2-4d8b-b1fd-19e31fb35b77/1eebb320-30e2-4d8b-b1fd-19e31fb35b77.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 799.952114] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-407a09ab-2e38-4fd8-9ea0-91526d1c68d1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.956551] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 799.956551] env[65522]: value = "task-5113986" [ 799.956551] env[65522]: _type = "Task" [ 799.956551] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.967217] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113986, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.151917] env[65522]: WARNING neutronclient.v2_0.client [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 800.152624] env[65522]: WARNING openstack [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 800.152966] env[65522]: WARNING openstack [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 800.174974] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521c87dc-8f6e-8abe-8cdf-e8cced9fca18, 'name': SearchDatastore_Task, 'duration_secs': 0.016147} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.175729] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 800.176047] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 800.176258] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.176409] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 800.176659] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 800.176880] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b067d0fa-de79-4645-9a70-1ad6a4c9fcbe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.192303] env[65522]: DEBUG nova.network.neutron [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Successfully updated port: 4c16f1be-de7b-45f4-9510-b1e2d358c37c {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 800.196240] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 800.196240] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 800.202583] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac0c9b19-60c9-45e5-a5ce-b071fac84860 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.211253] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Waiting for the task: (returnval){ [ 800.211253] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cbd5f3-0aaf-c2d8-3049-82225e7c7188" [ 800.211253] env[65522]: _type = "Task" [ 800.211253] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.215677] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5113985, 'name': ReconfigVM_Task, 'duration_secs': 0.713711} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.220806] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Reconfigured VM instance instance-0000002d to attach disk [datastore2] d2fdffb8-76d1-4b74-819f-25eb29ca91ce/d2fdffb8-76d1-4b74-819f-25eb29ca91ce.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.222206] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b998f17-d695-42b3-8847-3ebe9d15ddda {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.232589] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cbd5f3-0aaf-c2d8-3049-82225e7c7188, 'name': SearchDatastore_Task, 'duration_secs': 0.01826} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.236124] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Waiting for the task: (returnval){ [ 800.236124] env[65522]: value = "task-5113987" [ 800.236124] env[65522]: _type = "Task" [ 800.236124] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.237694] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eefba6c0-0279-4c78-90ca-1156148122c5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.248689] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Waiting for the task: (returnval){ [ 800.248689] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f83eda-8d85-caff-45b5-8255b6179052" [ 800.248689] env[65522]: _type = "Task" [ 800.248689] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.252741] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5113987, 'name': Rename_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.272889] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "7617189c-a902-42e7-8165-0e7c4a1de06d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 800.273518] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "7617189c-a902-42e7-8165-0e7c4a1de06d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 800.274102] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f83eda-8d85-caff-45b5-8255b6179052, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.288520] env[65522]: INFO nova.compute.manager [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] instance snapshotting [ 800.292495] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5856b980-2df2-49e2-8f74-7f06e76a3dac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.316938] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc03aca2-c95f-4744-9b79-859b43ed8cd2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.406913] env[65522]: WARNING openstack [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 800.408334] env[65522]: WARNING openstack [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 800.429507] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.891s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 800.431847] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.922s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 800.432108] env[65522]: DEBUG nova.objects.instance [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Lazy-loading 'resources' on Instance uuid 6e10e846-cdb7-458e-b511-18ab1742228e {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 800.470132] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113986, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.505676] env[65522]: WARNING neutronclient.v2_0.client [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 800.505676] env[65522]: WARNING openstack [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 800.506048] env[65522]: WARNING openstack [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 800.628102] env[65522]: DEBUG nova.network.neutron [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Updated VIF entry in instance network info cache for port e12c2307-4963-4a5a-af0f-19369217f430. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 800.628573] env[65522]: DEBUG nova.network.neutron [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Updating instance_info_cache with network_info: [{"id": "e12c2307-4963-4a5a-af0f-19369217f430", "address": "fa:16:3e:9f:b8:7d", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape12c2307-49", "ovs_interfaceid": "e12c2307-4963-4a5a-af0f-19369217f430", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 800.698702] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Acquiring lock "refresh_cache-cdee3a32-ad51-4364-8440-4744e8df322f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.701078] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Acquired lock "refresh_cache-cdee3a32-ad51-4364-8440-4744e8df322f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 800.701078] env[65522]: DEBUG nova.network.neutron [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 800.743792] env[65522]: DEBUG nova.network.neutron [-] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 800.753476] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5113987, 'name': Rename_Task, 'duration_secs': 0.163651} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.760643] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 800.760643] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c352bbbc-349c-42dc-a98d-79ed5b5a93ae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.769352] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f83eda-8d85-caff-45b5-8255b6179052, 'name': SearchDatastore_Task, 'duration_secs': 0.015533} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.770995] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 800.771306] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 9fa6e208-f913-4ecc-8548-ed6503fe6273/9fa6e208-f913-4ecc-8548-ed6503fe6273.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 800.771700] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Waiting for the task: (returnval){ [ 800.771700] env[65522]: value = "task-5113988" [ 800.771700] env[65522]: _type = "Task" [ 800.771700] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.771919] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a4e35c17-b923-405e-990e-8c76d6a648b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.787110] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Waiting for the task: (returnval){ [ 800.787110] env[65522]: value = "task-5113989" [ 800.787110] env[65522]: _type = "Task" [ 800.787110] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.797628] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5113989, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.843829] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 800.844796] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a23a759e-17f7-49f7-968c-3eb5a31367cc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.855089] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 800.855089] env[65522]: value = "task-5113990" [ 800.855089] env[65522]: _type = "Task" [ 800.855089] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.866590] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113990, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.870816] env[65522]: DEBUG nova.compute.manager [req-76ee3db7-4ce5-49de-9fbe-1c1d011ed7fd req-a6387d88-37c8-4311-934c-1d0466c84429 service nova] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Received event network-vif-plugged-4c16f1be-de7b-45f4-9510-b1e2d358c37c {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 800.871012] env[65522]: DEBUG oslo_concurrency.lockutils [req-76ee3db7-4ce5-49de-9fbe-1c1d011ed7fd req-a6387d88-37c8-4311-934c-1d0466c84429 service nova] Acquiring lock "cdee3a32-ad51-4364-8440-4744e8df322f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 800.872733] env[65522]: DEBUG oslo_concurrency.lockutils [req-76ee3db7-4ce5-49de-9fbe-1c1d011ed7fd req-a6387d88-37c8-4311-934c-1d0466c84429 service nova] Lock "cdee3a32-ad51-4364-8440-4744e8df322f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 800.873256] env[65522]: DEBUG oslo_concurrency.lockutils [req-76ee3db7-4ce5-49de-9fbe-1c1d011ed7fd req-a6387d88-37c8-4311-934c-1d0466c84429 service nova] Lock "cdee3a32-ad51-4364-8440-4744e8df322f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 800.873384] env[65522]: DEBUG nova.compute.manager [req-76ee3db7-4ce5-49de-9fbe-1c1d011ed7fd req-a6387d88-37c8-4311-934c-1d0466c84429 service nova] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] No waiting events found dispatching network-vif-plugged-4c16f1be-de7b-45f4-9510-b1e2d358c37c {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 800.873572] env[65522]: WARNING nova.compute.manager [req-76ee3db7-4ce5-49de-9fbe-1c1d011ed7fd req-a6387d88-37c8-4311-934c-1d0466c84429 service nova] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Received unexpected event network-vif-plugged-4c16f1be-de7b-45f4-9510-b1e2d358c37c for instance with vm_state building and task_state spawning. [ 800.972613] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113986, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.004759] env[65522]: INFO nova.scheduler.client.report [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Deleted allocation for migration 86317211-4746-4c8e-8603-9fd37855968e [ 801.117035] env[65522]: DEBUG nova.compute.manager [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 801.117824] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9329fc86-2ab6-48f4-a9de-0925e58af0a9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.133523] env[65522]: DEBUG oslo_concurrency.lockutils [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Releasing lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 801.133797] env[65522]: DEBUG nova.compute.manager [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Received event network-vif-plugged-7e072218-2d88-437a-b6fb-597ced883532 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 801.134017] env[65522]: DEBUG oslo_concurrency.lockutils [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Acquiring lock "9fa6e208-f913-4ecc-8548-ed6503fe6273-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 801.134224] env[65522]: DEBUG oslo_concurrency.lockutils [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Lock "9fa6e208-f913-4ecc-8548-ed6503fe6273-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 801.134369] env[65522]: DEBUG oslo_concurrency.lockutils [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Lock "9fa6e208-f913-4ecc-8548-ed6503fe6273-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 801.134539] env[65522]: DEBUG nova.compute.manager [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] No waiting events found dispatching network-vif-plugged-7e072218-2d88-437a-b6fb-597ced883532 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 801.134703] env[65522]: WARNING nova.compute.manager [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Received unexpected event network-vif-plugged-7e072218-2d88-437a-b6fb-597ced883532 for instance with vm_state building and task_state spawning. [ 801.134872] env[65522]: DEBUG nova.compute.manager [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Received event network-changed-7e072218-2d88-437a-b6fb-597ced883532 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 801.135047] env[65522]: DEBUG nova.compute.manager [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Refreshing instance network info cache due to event network-changed-7e072218-2d88-437a-b6fb-597ced883532. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 801.135232] env[65522]: DEBUG oslo_concurrency.lockutils [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Acquiring lock "refresh_cache-9fa6e208-f913-4ecc-8548-ed6503fe6273" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.135374] env[65522]: DEBUG oslo_concurrency.lockutils [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Acquired lock "refresh_cache-9fa6e208-f913-4ecc-8548-ed6503fe6273" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 801.135719] env[65522]: DEBUG nova.network.neutron [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Refreshing network info cache for port 7e072218-2d88-437a-b6fb-597ced883532 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 801.203907] env[65522]: WARNING openstack [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 801.204346] env[65522]: WARNING openstack [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 801.248557] env[65522]: INFO nova.compute.manager [-] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Took 1.62 seconds to deallocate network for instance. [ 801.258516] env[65522]: DEBUG nova.network.neutron [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 801.288627] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5113988, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.294724] env[65522]: WARNING openstack [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 801.295381] env[65522]: WARNING openstack [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 801.315927] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5113989, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.351461] env[65522]: DEBUG nova.compute.manager [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Stashing vm_state: active {{(pid=65522) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 801.370220] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113990, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.385887] env[65522]: WARNING neutronclient.v2_0.client [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 801.386606] env[65522]: WARNING openstack [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 801.386961] env[65522]: WARNING openstack [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 801.472838] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113986, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.489658] env[65522]: DEBUG nova.network.neutron [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Updating instance_info_cache with network_info: [{"id": "4c16f1be-de7b-45f4-9510-b1e2d358c37c", "address": "fa:16:3e:15:a6:4d", "network": {"id": "9f8e0708-f24b-4cbd-b89d-039f4d2d4f35", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1740139904-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecf042e8e8245cdb7e58d41b80e6737", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1316f5aa-529f-4bac-8dd7-6076a9d43312", "external-id": "nsx-vlan-transportzone-399", "segmentation_id": 399, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c16f1be-de", "ovs_interfaceid": "4c16f1be-de7b-45f4-9510-b1e2d358c37c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 801.510812] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15c2d7be-99be-4403-8a1d-c1104721cce9 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 30.837s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 801.605276] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23aec63b-d0fb-462a-a6f2-8f5e2c931e6d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.613971] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d359e55-fa88-4f2b-b09e-dc231cdecad4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.649566] env[65522]: INFO nova.compute.manager [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] instance snapshotting [ 801.651494] env[65522]: WARNING neutronclient.v2_0.client [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 801.652204] env[65522]: WARNING openstack [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 801.652552] env[65522]: WARNING openstack [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 801.664492] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43297b96-1b7b-42eb-92d5-964d533daa61 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.667752] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec3625c-6842-4c87-bf87-dc0664432867 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.700361] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6ab711-1131-4c7f-b844-c0d49eafc7c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.704830] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f057933-ff55-407b-8a69-ff3873d8e0a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.722232] env[65522]: DEBUG nova.compute.provider_tree [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.764357] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 801.800534] env[65522]: DEBUG oslo_vmware.api [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5113988, 'name': PowerOnVM_Task, 'duration_secs': 0.558757} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.801884] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 801.802279] env[65522]: INFO nova.compute.manager [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Took 10.62 seconds to spawn the instance on the hypervisor. [ 801.802706] env[65522]: DEBUG nova.compute.manager [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 801.805600] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b7e0b1-dba4-4969-ae3e-a7f0b5d5c468 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.816171] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5113989, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.849868] env[65522]: WARNING openstack [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 801.850505] env[65522]: WARNING openstack [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 801.868774] env[65522]: DEBUG nova.compute.manager [req-4ec45ae4-87e8-4c3d-b086-4551eda2644a req-a6edea45-ecfe-4baf-ad26-4162a6e2b4e5 service nova] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Received event network-vif-deleted-d24163e8-4f08-4908-9307-c4edb0d4ffb7 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 801.875750] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113990, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.890234] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 801.932779] env[65522]: WARNING neutronclient.v2_0.client [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 801.933495] env[65522]: WARNING openstack [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 801.933841] env[65522]: WARNING openstack [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 801.971816] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113986, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.783867} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.972094] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 1eebb320-30e2-4d8b-b1fd-19e31fb35b77/1eebb320-30e2-4d8b-b1fd-19e31fb35b77.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 801.972314] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.972617] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-593fd14a-3e0d-49b1-bd7f-2ca6175b8821 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.983746] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 801.983746] env[65522]: value = "task-5113991" [ 801.983746] env[65522]: _type = "Task" [ 801.983746] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.998158] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Releasing lock "refresh_cache-cdee3a32-ad51-4364-8440-4744e8df322f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 802.002026] env[65522]: DEBUG nova.compute.manager [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Instance network_info: |[{"id": "4c16f1be-de7b-45f4-9510-b1e2d358c37c", "address": "fa:16:3e:15:a6:4d", "network": {"id": "9f8e0708-f24b-4cbd-b89d-039f4d2d4f35", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1740139904-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecf042e8e8245cdb7e58d41b80e6737", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1316f5aa-529f-4bac-8dd7-6076a9d43312", "external-id": "nsx-vlan-transportzone-399", "segmentation_id": 399, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c16f1be-de", "ovs_interfaceid": "4c16f1be-de7b-45f4-9510-b1e2d358c37c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 802.002026] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113991, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.002026] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:a6:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1316f5aa-529f-4bac-8dd7-6076a9d43312', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c16f1be-de7b-45f4-9510-b1e2d358c37c', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 802.012561] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Creating folder: Project (6ecf042e8e8245cdb7e58d41b80e6737). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.020810] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-775d44b0-f1c5-4563-a5d3-d1a9f9804300 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.035304] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Created folder: Project (6ecf042e8e8245cdb7e58d41b80e6737) in parent group-v994660. [ 802.035604] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Creating folder: Instances. Parent ref: group-v994787. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.035998] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-97a8c113-95d8-473a-9ffb-d7b2cd7b101d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.048744] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Created folder: Instances in parent group-v994787. [ 802.049928] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 802.049928] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 802.049928] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-851c8e65-6348-42fc-bc16-3651d94eadca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.072342] env[65522]: DEBUG nova.network.neutron [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Updated VIF entry in instance network info cache for port 7e072218-2d88-437a-b6fb-597ced883532. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 802.072756] env[65522]: DEBUG nova.network.neutron [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Updating instance_info_cache with network_info: [{"id": "7e072218-2d88-437a-b6fb-597ced883532", "address": "fa:16:3e:21:b1:bf", "network": {"id": "0ab7f06c-0131-43cd-8161-2b87ea9cfb91", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-160966193-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "6ab0a321a59248a286c9d8775c7a9ddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e072218-2d", "ovs_interfaceid": "7e072218-2d88-437a-b6fb-597ced883532", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 802.075400] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 802.075400] env[65522]: value = "task-5113994" [ 802.075400] env[65522]: _type = "Task" [ 802.075400] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.086221] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113994, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.176989] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "cd82f2fd-d657-4ece-8e90-9a47072f5546" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 802.176989] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "cd82f2fd-d657-4ece-8e90-9a47072f5546" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 802.176989] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "cd82f2fd-d657-4ece-8e90-9a47072f5546-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 802.177197] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "cd82f2fd-d657-4ece-8e90-9a47072f5546-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 802.177238] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "cd82f2fd-d657-4ece-8e90-9a47072f5546-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 802.180292] env[65522]: INFO nova.compute.manager [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Terminating instance [ 802.231700] env[65522]: DEBUG nova.scheduler.client.report [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 802.237459] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 802.238115] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-583c26d1-7bc2-4449-b04b-2450685d396d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.248019] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 802.248019] env[65522]: value = "task-5113995" [ 802.248019] env[65522]: _type = "Task" [ 802.248019] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.262513] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113995, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.302847] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5113989, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.334889] env[65522]: INFO nova.compute.manager [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Took 32.19 seconds to build instance. [ 802.377339] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113990, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.496595] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113991, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096281} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.496740] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 802.498284] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d111a5f4-944f-4ed2-a049-1efa8748dccc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.536545] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] 1eebb320-30e2-4d8b-b1fd-19e31fb35b77/1eebb320-30e2-4d8b-b1fd-19e31fb35b77.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 802.536917] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-599d413d-b5ca-4bde-b972-598145a40e23 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.571603] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 802.571603] env[65522]: value = "task-5113996" [ 802.571603] env[65522]: _type = "Task" [ 802.571603] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.577380] env[65522]: DEBUG oslo_concurrency.lockutils [req-03728368-c41a-4c36-bbe8-6912d9e68be0 req-f387f73a-b008-420e-9539-c0ac8ad024dc service nova] Releasing lock "refresh_cache-9fa6e208-f913-4ecc-8548-ed6503fe6273" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 802.584891] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113996, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.590747] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113994, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.686061] env[65522]: DEBUG nova.compute.manager [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 802.686447] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 802.688319] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c67e2d-0136-45b4-bfc9-1398a9885100 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.698804] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 802.699605] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f1db8b8-2d83-4e2e-b217-a868c1b0ceaa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.708154] env[65522]: DEBUG oslo_vmware.api [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 802.708154] env[65522]: value = "task-5113997" [ 802.708154] env[65522]: _type = "Task" [ 802.708154] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.720217] env[65522]: DEBUG oslo_vmware.api [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113997, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.738995] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.307s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 802.742012] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.185s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 802.744041] env[65522]: INFO nova.compute.claims [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.761795] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113995, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.780317] env[65522]: INFO nova.scheduler.client.report [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Deleted allocations for instance 6e10e846-cdb7-458e-b511-18ab1742228e [ 802.806050] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5113989, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.522413} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.806050] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 9fa6e208-f913-4ecc-8548-ed6503fe6273/9fa6e208-f913-4ecc-8548-ed6503fe6273.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 802.806050] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 802.806506] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19124c67-49c0-49a2-b7bb-67abee0f8943 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.815855] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Waiting for the task: (returnval){ [ 802.815855] env[65522]: value = "task-5113998" [ 802.815855] env[65522]: _type = "Task" [ 802.815855] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.837440] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5113998, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.839470] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6d11d95-ea72-4da7-8bcd-6270c1afd4fe tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Lock "d2fdffb8-76d1-4b74-819f-25eb29ca91ce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.704s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 802.878276] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113990, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.966199] env[65522]: DEBUG nova.compute.manager [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Received event network-changed-4c16f1be-de7b-45f4-9510-b1e2d358c37c {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 802.967012] env[65522]: DEBUG nova.compute.manager [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Refreshing instance network info cache due to event network-changed-4c16f1be-de7b-45f4-9510-b1e2d358c37c. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 802.967012] env[65522]: DEBUG oslo_concurrency.lockutils [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] Acquiring lock "refresh_cache-cdee3a32-ad51-4364-8440-4744e8df322f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.967012] env[65522]: DEBUG oslo_concurrency.lockutils [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] Acquired lock "refresh_cache-cdee3a32-ad51-4364-8440-4744e8df322f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 802.967012] env[65522]: DEBUG nova.network.neutron [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Refreshing network info cache for port 4c16f1be-de7b-45f4-9510-b1e2d358c37c {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 803.090483] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113996, 'name': ReconfigVM_Task, 'duration_secs': 0.394791} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.091186] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Reconfigured VM instance instance-0000002e to attach disk [datastore2] 1eebb320-30e2-4d8b-b1fd-19e31fb35b77/1eebb320-30e2-4d8b-b1fd-19e31fb35b77.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 803.091977] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5973a9a9-0671-42d1-aea8-11555c53c7f8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.097807] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5113994, 'name': CreateVM_Task, 'duration_secs': 0.528959} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.098505] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 803.099380] env[65522]: WARNING neutronclient.v2_0.client [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 803.099783] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.099990] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 803.100365] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 803.100664] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a35882d0-f62f-40ee-8e10-51be7ca0025d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.105653] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 803.105653] env[65522]: value = "task-5113999" [ 803.105653] env[65522]: _type = "Task" [ 803.105653] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.110353] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Waiting for the task: (returnval){ [ 803.110353] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523d64bd-510b-8819-b4dc-91e0b1020615" [ 803.110353] env[65522]: _type = "Task" [ 803.110353] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.117627] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113999, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.125293] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523d64bd-510b-8819-b4dc-91e0b1020615, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.185607] env[65522]: DEBUG oslo_concurrency.lockutils [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 803.186112] env[65522]: DEBUG oslo_concurrency.lockutils [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 803.186420] env[65522]: DEBUG oslo_concurrency.lockutils [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 803.186760] env[65522]: DEBUG oslo_concurrency.lockutils [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 803.187063] env[65522]: DEBUG oslo_concurrency.lockutils [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 803.191425] env[65522]: INFO nova.compute.manager [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Terminating instance [ 803.220448] env[65522]: DEBUG oslo_vmware.api [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5113997, 'name': PowerOffVM_Task, 'duration_secs': 0.233316} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.220843] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 803.221544] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 803.221544] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eac160e2-be08-43e1-bd31-ba9b4e21571f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.262604] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113995, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.293322] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 803.293520] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 803.294734] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Deleting the datastore file [datastore2] cd82f2fd-d657-4ece-8e90-9a47072f5546 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 803.294734] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d8d9a8da-f623-43df-9aec-1c91aa05d7c4 tempest-ServersAdmin275Test-904595622 tempest-ServersAdmin275Test-904595622-project-member] Lock "6e10e846-cdb7-458e-b511-18ab1742228e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.066s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 803.295541] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f00f877-8ee3-42ac-a8fb-95ad46ae6c92 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.307983] env[65522]: DEBUG oslo_vmware.api [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 803.307983] env[65522]: value = "task-5114001" [ 803.307983] env[65522]: _type = "Task" [ 803.307983] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.319926] env[65522]: DEBUG oslo_vmware.api [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5114001, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.331923] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5113998, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08763} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.332416] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 803.334458] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db93d0d-a9dd-43f0-9055-c6abeaa8f3b3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.351562] env[65522]: DEBUG nova.compute.manager [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 803.364479] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] 9fa6e208-f913-4ecc-8548-ed6503fe6273/9fa6e208-f913-4ecc-8548-ed6503fe6273.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.365715] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3921534d-c218-43f0-a49f-21c4a623ca76 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.395601] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5113990, 'name': CreateSnapshot_Task, 'duration_secs': 2.36251} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.395926] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 803.396278] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Waiting for the task: (returnval){ [ 803.396278] env[65522]: value = "task-5114002" [ 803.396278] env[65522]: _type = "Task" [ 803.396278] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.396943] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-198798f8-5760-410c-8f6b-0a16f6021596 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.471427] env[65522]: WARNING neutronclient.v2_0.client [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 803.472200] env[65522]: WARNING openstack [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 803.472656] env[65522]: WARNING openstack [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 803.587821] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "028c7fc5-5624-4dd5-9e2f-48191a86f765" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 803.588076] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "028c7fc5-5624-4dd5-9e2f-48191a86f765" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 803.588267] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "028c7fc5-5624-4dd5-9e2f-48191a86f765-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 803.588453] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "028c7fc5-5624-4dd5-9e2f-48191a86f765-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 803.588616] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "028c7fc5-5624-4dd5-9e2f-48191a86f765-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 803.591110] env[65522]: INFO nova.compute.manager [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Terminating instance [ 803.617264] env[65522]: WARNING openstack [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 803.617692] env[65522]: WARNING openstack [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 803.629862] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5113999, 'name': Rename_Task, 'duration_secs': 0.155134} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.630153] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 803.630435] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e90020b-094f-47c4-b904-fe3f1620a4ba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.637242] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523d64bd-510b-8819-b4dc-91e0b1020615, 'name': SearchDatastore_Task, 'duration_secs': 0.015408} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.638107] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 803.638473] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 803.638839] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.639102] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 803.639441] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 803.639851] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-810c04a6-944e-40d9-9b4d-5ac36f6ba48a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.643996] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 803.643996] env[65522]: value = "task-5114003" [ 803.643996] env[65522]: _type = "Task" [ 803.643996] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.658782] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 803.659010] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 803.667048] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e3d3df8-072a-4c6d-8fe4-11ce52d0a119 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.668460] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114003, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.673047] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Waiting for the task: (returnval){ [ 803.673047] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b458fa-b706-755a-5290-f90019ddce1b" [ 803.673047] env[65522]: _type = "Task" [ 803.673047] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.684510] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b458fa-b706-755a-5290-f90019ddce1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.687289] env[65522]: WARNING neutronclient.v2_0.client [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 803.688010] env[65522]: WARNING openstack [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 803.688586] env[65522]: WARNING openstack [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 803.697888] env[65522]: DEBUG nova.compute.manager [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 803.698138] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 803.699351] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5eb573-66f3-46f2-a079-7d89138d5775 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.707877] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 803.707877] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-288a6eaa-3873-4a3b-b5bd-0ddcdb772c3a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.716613] env[65522]: DEBUG oslo_vmware.api [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 803.716613] env[65522]: value = "task-5114004" [ 803.716613] env[65522]: _type = "Task" [ 803.716613] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.727572] env[65522]: DEBUG oslo_vmware.api [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5114004, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.763849] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5113995, 'name': CreateSnapshot_Task, 'duration_secs': 1.254997} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.764411] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 803.765514] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78125971-2410-4c66-beac-ab98dde41ffe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.821929] env[65522]: DEBUG oslo_vmware.api [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5114001, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.227404} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.822299] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 803.823076] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 803.823076] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 803.823076] env[65522]: INFO nova.compute.manager [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Took 1.14 seconds to destroy the instance on the hypervisor. [ 803.823076] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 803.823517] env[65522]: DEBUG nova.compute.manager [-] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 803.823517] env[65522]: DEBUG nova.network.neutron [-] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 803.823579] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 803.825033] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 803.825033] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 803.833668] env[65522]: DEBUG nova.network.neutron [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Updated VIF entry in instance network info cache for port 4c16f1be-de7b-45f4-9510-b1e2d358c37c. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 803.834542] env[65522]: DEBUG nova.network.neutron [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Updating instance_info_cache with network_info: [{"id": "4c16f1be-de7b-45f4-9510-b1e2d358c37c", "address": "fa:16:3e:15:a6:4d", "network": {"id": "9f8e0708-f24b-4cbd-b89d-039f4d2d4f35", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1740139904-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecf042e8e8245cdb7e58d41b80e6737", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1316f5aa-529f-4bac-8dd7-6076a9d43312", "external-id": "nsx-vlan-transportzone-399", "segmentation_id": 399, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c16f1be-de", "ovs_interfaceid": "4c16f1be-de7b-45f4-9510-b1e2d358c37c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 803.882667] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 803.891091] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 803.916244] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5114002, 'name': ReconfigVM_Task, 'duration_secs': 0.345174} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.916244] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Reconfigured VM instance instance-0000002f to attach disk [datastore2] 9fa6e208-f913-4ecc-8548-ed6503fe6273/9fa6e208-f913-4ecc-8548-ed6503fe6273.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 803.916424] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d585232-a046-4446-a8a8-fc19038e60cd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.925905] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 803.926237] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b2b967fa-2954-4b5c-8091-d53868b00009 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.941304] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Waiting for the task: (returnval){ [ 803.941304] env[65522]: value = "task-5114005" [ 803.941304] env[65522]: _type = "Task" [ 803.941304] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.946011] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 803.946011] env[65522]: value = "task-5114006" [ 803.946011] env[65522]: _type = "Task" [ 803.946011] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.955191] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5114005, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.964311] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114006, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.095372] env[65522]: DEBUG nova.compute.manager [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 804.095658] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 804.096937] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b6432e-5354-4a3e-972e-a114c28e1de0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.108289] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 804.108289] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-574eb7fe-aa80-4043-9c81-ca26c3b6e760 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.118684] env[65522]: DEBUG oslo_vmware.api [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 804.118684] env[65522]: value = "task-5114007" [ 804.118684] env[65522]: _type = "Task" [ 804.118684] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.133562] env[65522]: DEBUG oslo_vmware.api [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5114007, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.161195] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114003, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.189676] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b458fa-b706-755a-5290-f90019ddce1b, 'name': SearchDatastore_Task, 'duration_secs': 0.012062} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.190558] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96d9423d-71bf-433e-81b5-aca5f09e609c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.201475] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Waiting for the task: (returnval){ [ 804.201475] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525f5728-32fc-97b1-c810-6ebf77fa77c5" [ 804.201475] env[65522]: _type = "Task" [ 804.201475] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.213238] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525f5728-32fc-97b1-c810-6ebf77fa77c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.233267] env[65522]: DEBUG oslo_vmware.api [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5114004, 'name': PowerOffVM_Task, 'duration_secs': 0.244075} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.237673] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 804.237673] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 804.237673] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6f018d20-3a69-4274-add9-e88dce5b5064 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.294593] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 804.299783] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-462df4fd-4951-4b6e-9d9c-b601d4c3468c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.312314] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 804.312314] env[65522]: value = "task-5114009" [ 804.312314] env[65522]: _type = "Task" [ 804.312314] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.328909] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114009, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.340033] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 804.340310] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 804.340572] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Deleting the datastore file [datastore1] 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 804.341301] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eda412d5-5a9d-4a24-8484-0f6fd71aecc0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.344246] env[65522]: DEBUG oslo_concurrency.lockutils [req-ff434ff0-c999-4cb5-aa85-30cd3b81ddff req-920377ee-d859-4581-9f00-e60c5f51e070 service nova] Releasing lock "refresh_cache-cdee3a32-ad51-4364-8440-4744e8df322f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 804.352202] env[65522]: DEBUG oslo_vmware.api [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 804.352202] env[65522]: value = "task-5114010" [ 804.352202] env[65522]: _type = "Task" [ 804.352202] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.366036] env[65522]: DEBUG oslo_vmware.api [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5114010, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.461126] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5114005, 'name': Rename_Task, 'duration_secs': 0.232901} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.461575] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 804.462472] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd943020-941a-4336-9ac2-8fa6aa2bf3d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.470391] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114006, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.481774] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Waiting for the task: (returnval){ [ 804.481774] env[65522]: value = "task-5114011" [ 804.481774] env[65522]: _type = "Task" [ 804.481774] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.495610] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5114011, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.584683] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f380f567-5390-4e59-b704-6b6cabe6c4a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.598805] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c256f409-3bfa-4a40-9127-babadfe35e9e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.638779] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f571e2-dceb-4cb9-bc53-26be8e56b56d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.647482] env[65522]: DEBUG oslo_vmware.api [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5114007, 'name': PowerOffVM_Task, 'duration_secs': 0.249798} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.653308] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 804.653549] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 804.654034] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f1bbcca8-efa9-43c9-84cd-66b7d09a5471 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.656982] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2394b2-4ce9-4a8e-b779-f474227d9f8d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.668769] env[65522]: DEBUG oslo_vmware.api [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114003, 'name': PowerOnVM_Task, 'duration_secs': 0.66227} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.676950] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 804.677310] env[65522]: INFO nova.compute.manager [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Took 10.77 seconds to spawn the instance on the hypervisor. [ 804.678278] env[65522]: DEBUG nova.compute.manager [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 804.678278] env[65522]: DEBUG nova.compute.provider_tree [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.680859] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a295962-6326-42c2-b28f-c9f5b56bb24f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.715086] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525f5728-32fc-97b1-c810-6ebf77fa77c5, 'name': SearchDatastore_Task, 'duration_secs': 0.019353} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.715445] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 804.715698] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] cdee3a32-ad51-4364-8440-4744e8df322f/cdee3a32-ad51-4364-8440-4744e8df322f.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 804.716021] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b24516cc-d4f3-4cd8-80d9-62a602f13eeb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.726444] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Waiting for the task: (returnval){ [ 804.726444] env[65522]: value = "task-5114013" [ 804.726444] env[65522]: _type = "Task" [ 804.726444] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.736610] env[65522]: DEBUG nova.network.neutron [-] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 804.738477] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114013, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.743022] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 804.743022] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 804.743022] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Deleting the datastore file [datastore1] 028c7fc5-5624-4dd5-9e2f-48191a86f765 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 804.743022] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a649f33-bb13-40da-8766-7e48ac5fa352 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.753053] env[65522]: DEBUG oslo_vmware.api [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for the task: (returnval){ [ 804.753053] env[65522]: value = "task-5114014" [ 804.753053] env[65522]: _type = "Task" [ 804.753053] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.765849] env[65522]: DEBUG oslo_vmware.api [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5114014, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.827038] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114009, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.862274] env[65522]: DEBUG oslo_vmware.api [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5114010, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.419218} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.863537] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 804.864051] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 804.864255] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 804.864494] env[65522]: INFO nova.compute.manager [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Took 1.17 seconds to destroy the instance on the hypervisor. [ 804.864845] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 804.865376] env[65522]: DEBUG nova.compute.manager [-] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 804.865504] env[65522]: DEBUG nova.network.neutron [-] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 804.865810] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 804.866389] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 804.866642] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 804.963806] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114006, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.994707] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5114011, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.020723] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 805.182046] env[65522]: DEBUG nova.scheduler.client.report [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 805.207605] env[65522]: INFO nova.compute.manager [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Took 34.02 seconds to build instance. [ 805.242807] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114013, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.243564] env[65522]: INFO nova.compute.manager [-] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Took 1.42 seconds to deallocate network for instance. [ 805.268646] env[65522]: DEBUG oslo_vmware.api [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Task: {'id': task-5114014, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234574} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.268918] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 805.269120] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 805.269409] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 805.269604] env[65522]: INFO nova.compute.manager [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Took 1.17 seconds to destroy the instance on the hypervisor. [ 805.269867] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 805.270196] env[65522]: DEBUG nova.compute.manager [-] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 805.270268] env[65522]: DEBUG nova.network.neutron [-] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 805.270498] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 805.271156] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 805.272039] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 805.330982] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114009, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.391430] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 805.436569] env[65522]: DEBUG nova.compute.manager [req-329f0441-e0a5-4789-a1ea-74d049b14316 req-a67a37e8-497c-49bc-817e-1513dbb8c47b service nova] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Received event network-vif-deleted-ff1e8448-0d72-4738-ae2b-9b543dbe9f5e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 805.473319] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114006, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.501235] env[65522]: DEBUG oslo_vmware.api [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5114011, 'name': PowerOnVM_Task, 'duration_secs': 0.68822} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.501514] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 805.501718] env[65522]: INFO nova.compute.manager [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Took 8.83 seconds to spawn the instance on the hypervisor. [ 805.502025] env[65522]: DEBUG nova.compute.manager [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 805.502839] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0457d9d9-f2d7-4423-801d-61b735bbb560 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.693066] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.948s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 805.693066] env[65522]: DEBUG nova.compute.manager [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 805.694686] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.651s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 805.695053] env[65522]: DEBUG nova.objects.instance [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lazy-loading 'resources' on Instance uuid f44db343-7e17-4780-9a67-245992ac1a63 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 805.710521] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a6fef3c-339a-472b-a03c-3f7fa1bb1489 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.553s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 805.739256] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114013, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.693585} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.739256] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] cdee3a32-ad51-4364-8440-4744e8df322f/cdee3a32-ad51-4364-8440-4744e8df322f.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 805.743019] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 805.743019] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c86a4ab3-11cb-45ef-9d37-f8f8c4733262 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.748837] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Waiting for the task: (returnval){ [ 805.748837] env[65522]: value = "task-5114015" [ 805.748837] env[65522]: _type = "Task" [ 805.748837] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.751679] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 805.763707] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114015, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.830149] env[65522]: DEBUG nova.network.neutron [-] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 805.830149] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114009, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.967049] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114006, 'name': CloneVM_Task} progress is 95%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.026873] env[65522]: INFO nova.compute.manager [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Took 31.57 seconds to build instance. [ 806.136023] env[65522]: DEBUG nova.network.neutron [-] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 806.139276] env[65522]: DEBUG nova.compute.manager [req-2f65c841-2858-43e8-b10d-1f7c0369e476 req-b4a40b4d-5d6d-4302-b189-d089962e1a3f service nova] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Received event network-vif-deleted-95b7c754-4f56-4bea-84f7-f4269d730b51 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 806.201573] env[65522]: DEBUG nova.compute.utils [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 806.205790] env[65522]: DEBUG nova.compute.manager [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Not allocating networking since 'none' was specified. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 806.266030] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114015, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094684} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.272348] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 806.274279] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499cac63-0766-4623-976e-a6323403ce4f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.305809] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] cdee3a32-ad51-4364-8440-4744e8df322f/cdee3a32-ad51-4364-8440-4744e8df322f.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 806.309298] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbca8ff8-eb7e-484e-9ad1-1bb88d683ed6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.338564] env[65522]: INFO nova.compute.manager [-] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Took 1.47 seconds to deallocate network for instance. [ 806.339103] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114009, 'name': CloneVM_Task} progress is 95%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.343764] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Waiting for the task: (returnval){ [ 806.343764] env[65522]: value = "task-5114016" [ 806.343764] env[65522]: _type = "Task" [ 806.343764] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.472473] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114006, 'name': CloneVM_Task, 'duration_secs': 2.436118} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.474095] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Created linked-clone VM from snapshot [ 806.474095] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2d58db-e895-4856-b364-22d6ef0d4329 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.486461] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Uploading image 987d36cc-3327-435f-9eb6-6f994a08ebfa {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 806.525995] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 806.525995] env[65522]: value = "vm-994792" [ 806.525995] env[65522]: _type = "VirtualMachine" [ 806.525995] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 806.526370] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-fc22c4e8-7801-4fe4-84c7-43eccc66f602 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.531323] env[65522]: DEBUG oslo_concurrency.lockutils [None req-134ca0bf-fc7e-4332-bccf-3395e72ed232 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Lock "9fa6e208-f913-4ecc-8548-ed6503fe6273" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.086s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.538283] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lease: (returnval){ [ 806.538283] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e3a2ad-ed4e-49b7-c2f5-b0cbefa9b6ac" [ 806.538283] env[65522]: _type = "HttpNfcLease" [ 806.538283] env[65522]: } obtained for exporting VM: (result){ [ 806.538283] env[65522]: value = "vm-994792" [ 806.538283] env[65522]: _type = "VirtualMachine" [ 806.538283] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 806.538283] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the lease: (returnval){ [ 806.538283] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e3a2ad-ed4e-49b7-c2f5-b0cbefa9b6ac" [ 806.538283] env[65522]: _type = "HttpNfcLease" [ 806.538283] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 806.546643] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 806.546643] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e3a2ad-ed4e-49b7-c2f5-b0cbefa9b6ac" [ 806.546643] env[65522]: _type = "HttpNfcLease" [ 806.546643] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 806.641141] env[65522]: INFO nova.compute.manager [-] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Took 1.37 seconds to deallocate network for instance. [ 806.711158] env[65522]: DEBUG nova.compute.manager [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 806.839726] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114009, 'name': CloneVM_Task, 'duration_secs': 2.282452} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.839726] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Created linked-clone VM from snapshot [ 806.840941] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-216f0379-a404-40ac-9ed3-86f96bd69f68 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.853862] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Uploading image 4d208f09-feb1-4bf6-a9e6-91683d8c378a {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 806.862208] env[65522]: DEBUG oslo_concurrency.lockutils [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.869904] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114016, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.872295] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 806.872700] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-43c316b7-2ac9-4945-b61d-8c93bf70b795 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.883161] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 806.883161] env[65522]: value = "task-5114018" [ 806.883161] env[65522]: _type = "Task" [ 806.883161] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.898059] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114018, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.981526] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3177f6f6-dea0-4a12-a30f-c2abb5960357 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.988524] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68eaf718-ceb9-4740-9d68-9977497fd332 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.024338] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f525d02b-d853-43e2-b703-122f9eaebe9d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.034818] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67686ed5-0d2b-4793-8bab-90cec4b6bff0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.048121] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 807.048121] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e3a2ad-ed4e-49b7-c2f5-b0cbefa9b6ac" [ 807.048121] env[65522]: _type = "HttpNfcLease" [ 807.048121] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 807.059422] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 807.059422] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e3a2ad-ed4e-49b7-c2f5-b0cbefa9b6ac" [ 807.059422] env[65522]: _type = "HttpNfcLease" [ 807.059422] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 807.059422] env[65522]: DEBUG nova.compute.provider_tree [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.062651] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f68de5f-ef4e-4a39-978f-fa119f13afb1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.073581] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529936a3-c101-b946-9e53-8bfd69a815a2/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 807.073780] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529936a3-c101-b946-9e53-8bfd69a815a2/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 807.151294] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 807.178029] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2fb13f66-ece3-4081-8601-a470667d10cc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.191802] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquiring lock "26432ad6-610e-49c2-8699-f1e7e128e5b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 807.192156] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Lock "26432ad6-610e-49c2-8699-f1e7e128e5b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 807.369330] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114016, 'name': ReconfigVM_Task, 'duration_secs': 0.763836} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.369768] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Reconfigured VM instance instance-00000030 to attach disk [datastore1] cdee3a32-ad51-4364-8440-4744e8df322f/cdee3a32-ad51-4364-8440-4744e8df322f.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 807.373047] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6f5834b5-544f-47cf-bd77-3b47df0fc907 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.381075] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Waiting for the task: (returnval){ [ 807.381075] env[65522]: value = "task-5114019" [ 807.381075] env[65522]: _type = "Task" [ 807.381075] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.399134] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114019, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.404978] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114018, 'name': Destroy_Task, 'duration_secs': 0.384954} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.405151] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Destroyed the VM [ 807.406240] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 807.406240] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c5d68646-834c-4891-bd74-c25aed6a0b11 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.419809] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 807.419809] env[65522]: value = "task-5114020" [ 807.419809] env[65522]: _type = "Task" [ 807.419809] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.431784] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114020, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.563880] env[65522]: DEBUG nova.scheduler.client.report [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 807.695834] env[65522]: DEBUG nova.compute.manager [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 807.728283] env[65522]: DEBUG nova.compute.manager [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 807.755185] env[65522]: DEBUG nova.virt.hardware [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 807.755584] env[65522]: DEBUG nova.virt.hardware [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 807.756088] env[65522]: DEBUG nova.virt.hardware [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 807.756297] env[65522]: DEBUG nova.virt.hardware [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 807.756454] env[65522]: DEBUG nova.virt.hardware [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 807.756662] env[65522]: DEBUG nova.virt.hardware [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 807.756959] env[65522]: DEBUG nova.virt.hardware [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 807.757191] env[65522]: DEBUG nova.virt.hardware [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 807.757528] env[65522]: DEBUG nova.virt.hardware [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 807.757826] env[65522]: DEBUG nova.virt.hardware [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 807.758094] env[65522]: DEBUG nova.virt.hardware [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 807.759921] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222761ce-e2b7-449a-ae44-550d1cfd87ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.772976] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f28a2f0-54d5-4dbf-ac83-21de3e4c508c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.795116] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.803073] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Creating folder: Project (3ae929571fb5465cba5bedffbb6aa4c8). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.803768] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-abe3a3c6-5f8f-4823-b295-2159ee3bc8e9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.820444] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Created folder: Project (3ae929571fb5465cba5bedffbb6aa4c8) in parent group-v994660. [ 807.820655] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Creating folder: Instances. Parent ref: group-v994794. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.821018] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-92ee64a4-1acd-4de8-adbb-ddc01a88be75 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.835255] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Created folder: Instances in parent group-v994794. [ 807.835881] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 807.836336] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 807.836684] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f14bd16d-0a1a-4760-8967-264320835a89 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.861156] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.861156] env[65522]: value = "task-5114023" [ 807.861156] env[65522]: _type = "Task" [ 807.861156] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.879076] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114023, 'name': CreateVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.892044] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114019, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.917162] env[65522]: DEBUG nova.compute.manager [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Received event network-changed-ad290ac3-4edd-4dd7-b683-0f0283c2dcc4 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 807.917597] env[65522]: DEBUG nova.compute.manager [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Refreshing instance network info cache due to event network-changed-ad290ac3-4edd-4dd7-b683-0f0283c2dcc4. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 807.918396] env[65522]: DEBUG oslo_concurrency.lockutils [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Acquiring lock "refresh_cache-d2fdffb8-76d1-4b74-819f-25eb29ca91ce" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.918396] env[65522]: DEBUG oslo_concurrency.lockutils [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Acquired lock "refresh_cache-d2fdffb8-76d1-4b74-819f-25eb29ca91ce" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 807.918396] env[65522]: DEBUG nova.network.neutron [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Refreshing network info cache for port ad290ac3-4edd-4dd7-b683-0f0283c2dcc4 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 807.935791] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114020, 'name': RemoveSnapshot_Task} progress is 76%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.068950] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.374s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 808.072446] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.843s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 808.074622] env[65522]: INFO nova.compute.claims [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 808.118570] env[65522]: INFO nova.scheduler.client.report [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Deleted allocations for instance f44db343-7e17-4780-9a67-245992ac1a63 [ 808.224132] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 808.377092] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquiring lock "9fa6e208-f913-4ecc-8548-ed6503fe6273" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 808.377740] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Lock "9fa6e208-f913-4ecc-8548-ed6503fe6273" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 808.378248] env[65522]: INFO nova.compute.manager [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Rebooting instance [ 808.380098] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114023, 'name': CreateVM_Task, 'duration_secs': 0.367796} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.380714] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 808.386830] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.387091] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 808.389155] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 808.389155] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-416b5494-c583-4d4d-a3c9-27b4d55b3309 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.396737] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Waiting for the task: (returnval){ [ 808.396737] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cd7bba-5cdc-a638-5f6c-bb5a3dd72b56" [ 808.396737] env[65522]: _type = "Task" [ 808.396737] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.402829] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114019, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.416191] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cd7bba-5cdc-a638-5f6c-bb5a3dd72b56, 'name': SearchDatastore_Task, 'duration_secs': 0.013669} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.416770] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 808.417166] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 808.417564] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.417881] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 808.418268] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 808.418773] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-30ce0683-55cb-4b99-937b-4b90985e4eed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.427347] env[65522]: WARNING neutronclient.v2_0.client [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 808.427347] env[65522]: WARNING openstack [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.427347] env[65522]: WARNING openstack [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 808.440887] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 808.441284] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 808.442280] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcff0f05-6cc1-4e0c-a604-4bbffa806d73 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.451598] env[65522]: DEBUG oslo_vmware.api [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114020, 'name': RemoveSnapshot_Task, 'duration_secs': 0.836105} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.453231] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 808.458075] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Waiting for the task: (returnval){ [ 808.458075] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c027e9-c24f-241a-ad5b-de6f739f14bc" [ 808.458075] env[65522]: _type = "Task" [ 808.458075] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.468342] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c027e9-c24f-241a-ad5b-de6f739f14bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.603635] env[65522]: WARNING openstack [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.604227] env[65522]: WARNING openstack [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 808.631221] env[65522]: DEBUG oslo_concurrency.lockutils [None req-954121f1-072c-4ed9-a661-2787be998889 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "f44db343-7e17-4780-9a67-245992ac1a63" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.204s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 808.655721] env[65522]: DEBUG nova.compute.manager [req-92d3d082-1b4b-428c-a040-353239ea5e35 req-22210bca-df2b-4093-b457-f45a7b6ea82d service nova] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Received event network-vif-deleted-93f5a0f6-2499-4475-b6e9-68d105b143f3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 808.685540] env[65522]: WARNING neutronclient.v2_0.client [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 808.686275] env[65522]: WARNING openstack [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.687705] env[65522]: WARNING openstack [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 808.877151] env[65522]: DEBUG nova.network.neutron [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Updated VIF entry in instance network info cache for port ad290ac3-4edd-4dd7-b683-0f0283c2dcc4. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 808.878148] env[65522]: DEBUG nova.network.neutron [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Updating instance_info_cache with network_info: [{"id": "ad290ac3-4edd-4dd7-b683-0f0283c2dcc4", "address": "fa:16:3e:e9:2b:6d", "network": {"id": "2e4ac13b-587a-4147-96fe-a0dbacab7ebd", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1030195272-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04fdbd83ceb54e5fbbf8fe00aa66cb32", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad290ac3-4e", "ovs_interfaceid": "ad290ac3-4edd-4dd7-b683-0f0283c2dcc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 808.904996] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114019, 'name': Rename_Task, 'duration_secs': 1.207464} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.905328] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 808.905631] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5801cdb-d9fa-4a2d-ba73-6a8f2e73af26 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.917600] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Waiting for the task: (returnval){ [ 808.917600] env[65522]: value = "task-5114024" [ 808.917600] env[65522]: _type = "Task" [ 808.917600] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.921493] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquiring lock "refresh_cache-9fa6e208-f913-4ecc-8548-ed6503fe6273" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.921751] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquired lock "refresh_cache-9fa6e208-f913-4ecc-8548-ed6503fe6273" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 808.922020] env[65522]: DEBUG nova.network.neutron [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 808.932414] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114024, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.960951] env[65522]: WARNING nova.compute.manager [None req-a7cec93e-6bdf-4ed8-8036-e4b3b80a055a tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Image not found during snapshot: nova.exception.ImageNotFound: Image 4d208f09-feb1-4bf6-a9e6-91683d8c378a could not be found. [ 808.977667] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c027e9-c24f-241a-ad5b-de6f739f14bc, 'name': SearchDatastore_Task, 'duration_secs': 0.018307} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.978367] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e759a51-2f54-4f4f-95a0-0e54fb915fa9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.985954] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Waiting for the task: (returnval){ [ 808.985954] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52453ec6-9af2-bace-006a-5018ef92e0c6" [ 808.985954] env[65522]: _type = "Task" [ 808.985954] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.997769] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52453ec6-9af2-bace-006a-5018ef92e0c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.008775] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 809.008994] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.381431] env[65522]: DEBUG oslo_concurrency.lockutils [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Releasing lock "refresh_cache-d2fdffb8-76d1-4b74-819f-25eb29ca91ce" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 809.381786] env[65522]: DEBUG nova.compute.manager [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Received event network-changed-e12c2307-4963-4a5a-af0f-19369217f430 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 809.381786] env[65522]: DEBUG nova.compute.manager [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Refreshing instance network info cache due to event network-changed-e12c2307-4963-4a5a-af0f-19369217f430. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 809.382119] env[65522]: DEBUG oslo_concurrency.lockutils [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Acquiring lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.382387] env[65522]: DEBUG oslo_concurrency.lockutils [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Acquired lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 809.382677] env[65522]: DEBUG nova.network.neutron [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Refreshing network info cache for port e12c2307-4963-4a5a-af0f-19369217f430 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 809.433823] env[65522]: WARNING neutronclient.v2_0.client [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 809.434836] env[65522]: WARNING openstack [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.435447] env[65522]: WARNING openstack [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.444273] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114024, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.503976] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52453ec6-9af2-bace-006a-5018ef92e0c6, 'name': SearchDatastore_Task, 'duration_secs': 0.015555} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.503976] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 809.503976] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 1be995d0-3141-4dcd-ad9a-278d9874e457/1be995d0-3141-4dcd-ad9a-278d9874e457.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 809.503976] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6eb36d3f-7838-43ae-8254-21bac6f5630a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.513097] env[65522]: DEBUG nova.compute.manager [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 809.518553] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Waiting for the task: (returnval){ [ 809.518553] env[65522]: value = "task-5114025" [ 809.518553] env[65522]: _type = "Task" [ 809.518553] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.540605] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114025, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.617040] env[65522]: WARNING openstack [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.617538] env[65522]: WARNING openstack [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.674121] env[65522]: WARNING neutronclient.v2_0.client [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 809.674982] env[65522]: WARNING openstack [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.675419] env[65522]: WARNING openstack [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.789157] env[65522]: DEBUG nova.network.neutron [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Updating instance_info_cache with network_info: [{"id": "7e072218-2d88-437a-b6fb-597ced883532", "address": "fa:16:3e:21:b1:bf", "network": {"id": "0ab7f06c-0131-43cd-8161-2b87ea9cfb91", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-160966193-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "6ab0a321a59248a286c9d8775c7a9ddc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1eed7865-f9d8-463e-843f-3b0b3a962a2c", "external-id": "nsx-vlan-transportzone-852", "segmentation_id": 852, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e072218-2d", "ovs_interfaceid": "7e072218-2d88-437a-b6fb-597ced883532", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 809.804293] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e66a6b-90af-4e81-8268-dd1a17470e15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.815582] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce8be48-2355-4915-8b4d-066bc5300c67 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.855014] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499754d3-8c6d-4597-beb3-3a9c1407878b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.865812] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9c2747-2b32-4a0e-b59c-995931739234 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.885673] env[65522]: DEBUG nova.compute.provider_tree [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.891024] env[65522]: WARNING neutronclient.v2_0.client [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 809.891024] env[65522]: WARNING openstack [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.891024] env[65522]: WARNING openstack [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.903679] env[65522]: DEBUG nova.scheduler.client.report [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 809.927979] env[65522]: DEBUG oslo_vmware.api [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114024, 'name': PowerOnVM_Task, 'duration_secs': 0.70297} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.929088] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 809.929380] env[65522]: INFO nova.compute.manager [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Took 10.36 seconds to spawn the instance on the hypervisor. [ 809.929581] env[65522]: DEBUG nova.compute.manager [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 809.930590] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea9fd918-7756-45ba-b65f-d1216c34c082 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.040981] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114025, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.050215] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 810.092748] env[65522]: WARNING openstack [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.093317] env[65522]: WARNING openstack [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 810.193470] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "a8f9f223-023a-48fc-9186-a360a78e27a9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 810.194246] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "a8f9f223-023a-48fc-9186-a360a78e27a9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 810.194394] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "a8f9f223-023a-48fc-9186-a360a78e27a9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 810.195079] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "a8f9f223-023a-48fc-9186-a360a78e27a9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 810.195079] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "a8f9f223-023a-48fc-9186-a360a78e27a9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 810.198449] env[65522]: WARNING neutronclient.v2_0.client [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.199293] env[65522]: WARNING openstack [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.199656] env[65522]: WARNING openstack [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 810.210418] env[65522]: INFO nova.compute.manager [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Terminating instance [ 810.293233] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Releasing lock "refresh_cache-9fa6e208-f913-4ecc-8548-ed6503fe6273" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 810.324813] env[65522]: DEBUG nova.network.neutron [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Updated VIF entry in instance network info cache for port e12c2307-4963-4a5a-af0f-19369217f430. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 810.325154] env[65522]: DEBUG nova.network.neutron [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Updating instance_info_cache with network_info: [{"id": "e12c2307-4963-4a5a-af0f-19369217f430", "address": "fa:16:3e:9f:b8:7d", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape12c2307-49", "ovs_interfaceid": "e12c2307-4963-4a5a-af0f-19369217f430", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 810.412581] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 810.412581] env[65522]: DEBUG nova.compute.manager [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 810.415429] env[65522]: DEBUG oslo_concurrency.lockutils [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.650s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 810.415657] env[65522]: DEBUG nova.objects.instance [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lazy-loading 'resources' on Instance uuid 0d52b207-ac69-48ce-b3af-7a83a499886e {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 810.452857] env[65522]: INFO nova.compute.manager [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Took 35.85 seconds to build instance. [ 810.535840] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114025, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.716137} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.536155] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 1be995d0-3141-4dcd-ad9a-278d9874e457/1be995d0-3141-4dcd-ad9a-278d9874e457.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 810.536378] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 810.537130] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2a54fa44-db55-4416-8486-cce365cc05f2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.544935] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Waiting for the task: (returnval){ [ 810.544935] env[65522]: value = "task-5114026" [ 810.544935] env[65522]: _type = "Task" [ 810.544935] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.557885] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114026, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.715221] env[65522]: DEBUG nova.compute.manager [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 810.715527] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 810.716815] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f5145c-91ab-4a16-96aa-b410ed132389 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.727054] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 810.727416] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-effab29c-ec1e-4df6-aa99-88c74cf63d26 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.735793] env[65522]: DEBUG oslo_vmware.api [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 810.735793] env[65522]: value = "task-5114027" [ 810.735793] env[65522]: _type = "Task" [ 810.735793] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.745879] env[65522]: DEBUG oslo_vmware.api [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5114027, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.804679] env[65522]: DEBUG nova.compute.manager [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 810.804679] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731d8e16-716f-40d1-8ad8-85a8caaa7004 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.828241] env[65522]: DEBUG oslo_concurrency.lockutils [req-5d7a7b97-2580-4aab-a391-39b7e76d2ccd req-1fd5af33-53af-4ec3-a641-28229a152bdd service nova] Releasing lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 810.920477] env[65522]: DEBUG nova.compute.utils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 810.926072] env[65522]: DEBUG nova.compute.manager [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 810.926072] env[65522]: DEBUG nova.network.neutron [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 810.926072] env[65522]: WARNING neutronclient.v2_0.client [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.926072] env[65522]: WARNING neutronclient.v2_0.client [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.926072] env[65522]: WARNING openstack [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.926345] env[65522]: WARNING openstack [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 810.957700] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1318c4ac-9398-42e8-88f0-f7c5d9c5bad1 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Lock "cdee3a32-ad51-4364-8440-4744e8df322f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.439s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 810.991514] env[65522]: DEBUG nova.policy [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '149b8f2f034844f484632b8859ffd12b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '234a226c969b4965bfa2c411bb4c8d90', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 811.058162] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114026, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.125693} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.058473] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.060578] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b453273e-0e78-4455-b58d-67ce219341e1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.088978] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 1be995d0-3141-4dcd-ad9a-278d9874e457/1be995d0-3141-4dcd-ad9a-278d9874e457.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.092830] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6fabed52-1463-4cad-9951-097dd08414b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.108526] env[65522]: DEBUG oslo_concurrency.lockutils [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "7693f017-69b3-4ec4-ae3f-7e65ed34a471" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 811.108833] env[65522]: DEBUG oslo_concurrency.lockutils [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "7693f017-69b3-4ec4-ae3f-7e65ed34a471" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 811.109016] env[65522]: DEBUG oslo_concurrency.lockutils [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "7693f017-69b3-4ec4-ae3f-7e65ed34a471-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 811.109189] env[65522]: DEBUG oslo_concurrency.lockutils [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "7693f017-69b3-4ec4-ae3f-7e65ed34a471-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 811.109381] env[65522]: DEBUG oslo_concurrency.lockutils [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "7693f017-69b3-4ec4-ae3f-7e65ed34a471-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 811.112343] env[65522]: INFO nova.compute.manager [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Terminating instance [ 811.119030] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Waiting for the task: (returnval){ [ 811.119030] env[65522]: value = "task-5114028" [ 811.119030] env[65522]: _type = "Task" [ 811.119030] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.135252] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114028, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.136569] env[65522]: DEBUG nova.compute.manager [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Received event network-changed-ad290ac3-4edd-4dd7-b683-0f0283c2dcc4 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 811.136711] env[65522]: DEBUG nova.compute.manager [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Refreshing instance network info cache due to event network-changed-ad290ac3-4edd-4dd7-b683-0f0283c2dcc4. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 811.136907] env[65522]: DEBUG oslo_concurrency.lockutils [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] Acquiring lock "refresh_cache-d2fdffb8-76d1-4b74-819f-25eb29ca91ce" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.137445] env[65522]: DEBUG oslo_concurrency.lockutils [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] Acquired lock "refresh_cache-d2fdffb8-76d1-4b74-819f-25eb29ca91ce" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 811.137738] env[65522]: DEBUG nova.network.neutron [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Refreshing network info cache for port ad290ac3-4edd-4dd7-b683-0f0283c2dcc4 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 811.247062] env[65522]: DEBUG oslo_vmware.api [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5114027, 'name': PowerOffVM_Task, 'duration_secs': 0.292304} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.250112] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 811.250311] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 811.250841] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eeaba62e-1b24-48c0-9757-6a256e97f257 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.333857] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 811.334181] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 811.334710] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Deleting the datastore file [datastore1] a8f9f223-023a-48fc-9186-a360a78e27a9 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 811.335093] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ed4b0cb7-4744-46c3-8e5d-59b051bc63ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.347757] env[65522]: DEBUG oslo_vmware.api [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for the task: (returnval){ [ 811.347757] env[65522]: value = "task-5114030" [ 811.347757] env[65522]: _type = "Task" [ 811.347757] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.366027] env[65522]: DEBUG oslo_vmware.api [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5114030, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.374510] env[65522]: DEBUG nova.network.neutron [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Successfully created port: c7aacbb2-e620-48af-838a-6103f858f4c8 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 811.435948] env[65522]: DEBUG nova.compute.utils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 811.616926] env[65522]: DEBUG nova.compute.manager [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 811.617113] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 811.620924] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21c87d0-e89e-4d09-9673-7e8e98857eca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.635760] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114028, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.642509] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 811.643090] env[65522]: WARNING neutronclient.v2_0.client [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 811.643801] env[65522]: WARNING openstack [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 811.644223] env[65522]: WARNING openstack [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 811.651802] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e5b5d88-5914-4ba9-befd-b161ca5cbe57 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.660337] env[65522]: DEBUG oslo_vmware.api [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 811.660337] env[65522]: value = "task-5114031" [ 811.660337] env[65522]: _type = "Task" [ 811.660337] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.668374] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4726d6b7-3b5a-470a-b515-d927d5d5fd1c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.675124] env[65522]: DEBUG oslo_vmware.api [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114031, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.681377] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2931c4f8-e6e7-4757-8363-f4887d324374 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.718237] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a4623c-9ee8-4dfb-99a2-0998de1503a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.728329] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c16d82-d825-4089-a265-aa3ad714a478 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.743916] env[65522]: DEBUG nova.compute.provider_tree [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.808197] env[65522]: WARNING openstack [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 811.808620] env[65522]: WARNING openstack [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 811.835224] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257d6a7a-e3df-4b08-9055-6dd021fd3449 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.845098] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Doing hard reboot of VM {{(pid=65522) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 811.845411] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-2633e0f1-1b98-423c-b208-3569f3786388 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.862282] env[65522]: DEBUG oslo_vmware.api [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Waiting for the task: (returnval){ [ 811.862282] env[65522]: value = "task-5114032" [ 811.862282] env[65522]: _type = "Task" [ 811.862282] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.870541] env[65522]: DEBUG oslo_vmware.api [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Task: {'id': task-5114030, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.257891} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.871268] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 811.871386] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 811.871516] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 811.871807] env[65522]: INFO nova.compute.manager [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Took 1.16 seconds to destroy the instance on the hypervisor. [ 811.872079] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 811.872273] env[65522]: DEBUG nova.compute.manager [-] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 811.872364] env[65522]: DEBUG nova.network.neutron [-] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 811.872624] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 811.873195] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 811.873455] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 811.887788] env[65522]: DEBUG oslo_vmware.api [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5114032, 'name': ResetVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.916178] env[65522]: WARNING neutronclient.v2_0.client [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 811.917229] env[65522]: WARNING openstack [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 811.917229] env[65522]: WARNING openstack [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 811.943721] env[65522]: DEBUG nova.compute.manager [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 811.997867] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.138481] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114028, 'name': ReconfigVM_Task, 'duration_secs': 0.538027} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.138806] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 1be995d0-3141-4dcd-ad9a-278d9874e457/1be995d0-3141-4dcd-ad9a-278d9874e457.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.139725] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-76a43a23-deff-49c9-9fe0-2d67289d5a3b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.151145] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Waiting for the task: (returnval){ [ 812.151145] env[65522]: value = "task-5114033" [ 812.151145] env[65522]: _type = "Task" [ 812.151145] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.167122] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114033, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.178075] env[65522]: DEBUG oslo_vmware.api [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114031, 'name': PowerOffVM_Task, 'duration_secs': 0.265663} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.178449] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 812.178645] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 812.179559] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d60a687-be79-4428-9e9c-0d45d0d0129c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.247665] env[65522]: DEBUG nova.scheduler.client.report [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 812.288284] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 812.288607] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 812.288911] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Deleting the datastore file [datastore1] 7693f017-69b3-4ec4-ae3f-7e65ed34a471 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 812.289768] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8247d18e-c15d-4b80-8b54-c1b6ce6205b7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.300042] env[65522]: DEBUG oslo_vmware.api [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 812.300042] env[65522]: value = "task-5114035" [ 812.300042] env[65522]: _type = "Task" [ 812.300042] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.311717] env[65522]: DEBUG oslo_vmware.api [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114035, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.334868] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Acquiring lock "cdee3a32-ad51-4364-8440-4744e8df322f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 812.335496] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Lock "cdee3a32-ad51-4364-8440-4744e8df322f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.335837] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Acquiring lock "cdee3a32-ad51-4364-8440-4744e8df322f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 812.336151] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Lock "cdee3a32-ad51-4364-8440-4744e8df322f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.336467] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Lock "cdee3a32-ad51-4364-8440-4744e8df322f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.339870] env[65522]: INFO nova.compute.manager [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Terminating instance [ 812.382842] env[65522]: DEBUG oslo_vmware.api [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5114032, 'name': ResetVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.495066] env[65522]: DEBUG nova.network.neutron [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Updated VIF entry in instance network info cache for port ad290ac3-4edd-4dd7-b683-0f0283c2dcc4. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 812.497880] env[65522]: DEBUG nova.network.neutron [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Updating instance_info_cache with network_info: [{"id": "ad290ac3-4edd-4dd7-b683-0f0283c2dcc4", "address": "fa:16:3e:e9:2b:6d", "network": {"id": "2e4ac13b-587a-4147-96fe-a0dbacab7ebd", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1030195272-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04fdbd83ceb54e5fbbf8fe00aa66cb32", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad290ac3-4e", "ovs_interfaceid": "ad290ac3-4edd-4dd7-b683-0f0283c2dcc4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 812.665520] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114033, 'name': Rename_Task, 'duration_secs': 0.334443} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.666173] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 812.666294] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee77c1c2-7a1a-4ff0-ac28-883f90a994a4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.676688] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Waiting for the task: (returnval){ [ 812.676688] env[65522]: value = "task-5114036" [ 812.676688] env[65522]: _type = "Task" [ 812.676688] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.688454] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114036, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.725346] env[65522]: DEBUG oslo_concurrency.lockutils [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Acquiring lock "d2fdffb8-76d1-4b74-819f-25eb29ca91ce" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 812.725839] env[65522]: DEBUG oslo_concurrency.lockutils [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Lock "d2fdffb8-76d1-4b74-819f-25eb29ca91ce" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.726395] env[65522]: DEBUG oslo_concurrency.lockutils [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Acquiring lock "d2fdffb8-76d1-4b74-819f-25eb29ca91ce-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 812.726769] env[65522]: DEBUG oslo_concurrency.lockutils [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Lock "d2fdffb8-76d1-4b74-819f-25eb29ca91ce-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.727085] env[65522]: DEBUG oslo_concurrency.lockutils [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Lock "d2fdffb8-76d1-4b74-819f-25eb29ca91ce-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.730537] env[65522]: INFO nova.compute.manager [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Terminating instance [ 812.754538] env[65522]: DEBUG oslo_concurrency.lockutils [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.339s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.758937] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.347s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.759782] env[65522]: DEBUG nova.objects.instance [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Lazy-loading 'resources' on Instance uuid ea5b8e29-ebe6-444b-bd2d-029201003ee6 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.801877] env[65522]: INFO nova.scheduler.client.report [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Deleted allocations for instance 0d52b207-ac69-48ce-b3af-7a83a499886e [ 812.823166] env[65522]: DEBUG oslo_vmware.api [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114035, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.368243} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.823622] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 812.823984] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 812.824344] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 812.824579] env[65522]: INFO nova.compute.manager [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Took 1.21 seconds to destroy the instance on the hypervisor. [ 812.824975] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 812.825433] env[65522]: DEBUG nova.compute.manager [-] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 812.825578] env[65522]: DEBUG nova.network.neutron [-] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 812.825992] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.827027] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.827518] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.849218] env[65522]: DEBUG nova.compute.manager [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 812.849497] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 812.850887] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b52291b-89df-4604-a1b8-8ebb53fe071d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.865461] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 812.871132] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1fcdc004-2663-43fa-9e0d-ef3c8cae12c7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.886311] env[65522]: DEBUG oslo_vmware.api [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5114032, 'name': ResetVM_Task, 'duration_secs': 0.542172} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.888598] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Did hard reboot of VM {{(pid=65522) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 812.888963] env[65522]: DEBUG nova.compute.manager [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 812.889583] env[65522]: DEBUG oslo_vmware.api [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Waiting for the task: (returnval){ [ 812.889583] env[65522]: value = "task-5114037" [ 812.889583] env[65522]: _type = "Task" [ 812.889583] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.891113] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f11951d-9d4a-496c-8e2e-02257c51f18f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.902340] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.921613] env[65522]: DEBUG oslo_vmware.api [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.938612] env[65522]: DEBUG nova.network.neutron [-] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 812.959880] env[65522]: DEBUG nova.compute.manager [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 813.003207] env[65522]: DEBUG nova.virt.hardware [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:58:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=1,extra_specs={hw_rng:allowed='True'},flavorid='1070681115',id=19,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_1-1040669266',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 813.003659] env[65522]: DEBUG nova.virt.hardware [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 813.003945] env[65522]: DEBUG nova.virt.hardware [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 813.004326] env[65522]: DEBUG nova.virt.hardware [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 813.004580] env[65522]: DEBUG nova.virt.hardware [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 813.004874] env[65522]: DEBUG nova.virt.hardware [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 813.007259] env[65522]: DEBUG nova.virt.hardware [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 813.007490] env[65522]: DEBUG nova.virt.hardware [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 813.007947] env[65522]: DEBUG nova.virt.hardware [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 813.008295] env[65522]: DEBUG nova.virt.hardware [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 813.008511] env[65522]: DEBUG nova.virt.hardware [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 813.009126] env[65522]: DEBUG oslo_concurrency.lockutils [req-d5ceef3b-c280-470e-ae09-ccfed146d0e9 req-7c938b59-4c9a-4495-aba0-99a9ab62f8c7 service nova] Releasing lock "refresh_cache-d2fdffb8-76d1-4b74-819f-25eb29ca91ce" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 813.010929] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe80b63-d209-4543-af34-8a504ffdc458 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.022018] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71d948f-213b-4597-bb09-a8a01f45128e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.120630] env[65522]: DEBUG nova.network.neutron [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Successfully updated port: c7aacbb2-e620-48af-838a-6103f858f4c8 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 813.189238] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114036, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.235958] env[65522]: DEBUG nova.compute.manager [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 813.236348] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 813.237706] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bb11bf-e9b5-4e7f-b722-1954719cd74a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.249403] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 813.249843] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d06ad9c-60a1-4419-ad7c-4555792547ca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.258820] env[65522]: DEBUG oslo_vmware.api [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Waiting for the task: (returnval){ [ 813.258820] env[65522]: value = "task-5114038" [ 813.258820] env[65522]: _type = "Task" [ 813.258820] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.279209] env[65522]: DEBUG oslo_vmware.api [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5114038, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.317387] env[65522]: DEBUG oslo_concurrency.lockutils [None req-729f6605-05f6-4c34-8dd3-3809c89dc82d tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "0d52b207-ac69-48ce-b3af-7a83a499886e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.411s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 813.411711] env[65522]: DEBUG oslo_vmware.api [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114037, 'name': PowerOffVM_Task, 'duration_secs': 0.387948} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.415147] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 813.415437] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 813.416087] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4661feba-23fb-4cd0-b42c-e9278e0a5706 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.425985] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9a36a2e4-14b8-4ab1-aafa-70135b319e9b tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Lock "9fa6e208-f913-4ecc-8548-ed6503fe6273" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.048s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 813.443721] env[65522]: INFO nova.compute.manager [-] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Took 1.57 seconds to deallocate network for instance. [ 813.450967] env[65522]: DEBUG nova.compute.manager [req-85c9327d-8499-4d05-b9d2-0afd6a682228 req-699a1f8f-b351-485d-8cd7-bdf0d789bf3c service nova] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Received event network-vif-deleted-8cb15276-e492-4096-b65b-f1107c89c902 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 813.493971] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 813.493971] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 813.493971] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Deleting the datastore file [datastore1] cdee3a32-ad51-4364-8440-4744e8df322f {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 813.495772] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8ebc07f0-388e-4cd5-bfd8-7a21ff80bcb0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.513528] env[65522]: DEBUG oslo_vmware.api [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Waiting for the task: (returnval){ [ 813.513528] env[65522]: value = "task-5114040" [ 813.513528] env[65522]: _type = "Task" [ 813.513528] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.531476] env[65522]: DEBUG oslo_vmware.api [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114040, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.623452] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "refresh_cache-05319f69-2ac7-4649-a90a-3542d87a3e82" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.623720] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquired lock "refresh_cache-05319f69-2ac7-4649-a90a-3542d87a3e82" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 813.623912] env[65522]: DEBUG nova.network.neutron [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 813.689155] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114036, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.697462] env[65522]: DEBUG nova.network.neutron [-] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 813.762036] env[65522]: DEBUG nova.compute.manager [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Received event network-vif-plugged-c7aacbb2-e620-48af-838a-6103f858f4c8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 813.762365] env[65522]: DEBUG oslo_concurrency.lockutils [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] Acquiring lock "05319f69-2ac7-4649-a90a-3542d87a3e82-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 813.762660] env[65522]: DEBUG oslo_concurrency.lockutils [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] Lock "05319f69-2ac7-4649-a90a-3542d87a3e82-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 813.762892] env[65522]: DEBUG oslo_concurrency.lockutils [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] Lock "05319f69-2ac7-4649-a90a-3542d87a3e82-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 813.763153] env[65522]: DEBUG nova.compute.manager [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] No waiting events found dispatching network-vif-plugged-c7aacbb2-e620-48af-838a-6103f858f4c8 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 813.763464] env[65522]: WARNING nova.compute.manager [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Received unexpected event network-vif-plugged-c7aacbb2-e620-48af-838a-6103f858f4c8 for instance with vm_state building and task_state spawning. [ 813.763692] env[65522]: DEBUG nova.compute.manager [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Received event network-changed-c7aacbb2-e620-48af-838a-6103f858f4c8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 813.764031] env[65522]: DEBUG nova.compute.manager [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Refreshing instance network info cache due to event network-changed-c7aacbb2-e620-48af-838a-6103f858f4c8. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 813.764292] env[65522]: DEBUG oslo_concurrency.lockutils [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] Acquiring lock "refresh_cache-05319f69-2ac7-4649-a90a-3542d87a3e82" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.783321] env[65522]: DEBUG oslo_vmware.api [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5114038, 'name': PowerOffVM_Task, 'duration_secs': 0.303624} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.786819] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 813.786819] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 813.786819] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a094b7da-cf47-47fc-b050-03077b55eaa8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.873242] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 813.873242] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 813.873242] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Deleting the datastore file [datastore2] d2fdffb8-76d1-4b74-819f-25eb29ca91ce {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 813.874249] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79ab59e2-2225-423c-b7a3-36b8d933eb53 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.890293] env[65522]: DEBUG oslo_vmware.api [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Waiting for the task: (returnval){ [ 813.890293] env[65522]: value = "task-5114042" [ 813.890293] env[65522]: _type = "Task" [ 813.890293] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.911340] env[65522]: DEBUG oslo_vmware.api [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5114042, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.958181] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.008857] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b16f83-e151-4d48-847f-0d9ef069e373 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.021525] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a259950e-8f30-4b73-b438-b0cea22a9592 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.031839] env[65522]: DEBUG oslo_vmware.api [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Task: {'id': task-5114040, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.325859} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.059409] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 814.059659] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 814.059912] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 814.060122] env[65522]: INFO nova.compute.manager [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Took 1.21 seconds to destroy the instance on the hypervisor. [ 814.060437] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 814.061544] env[65522]: DEBUG nova.compute.manager [-] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 814.062027] env[65522]: DEBUG nova.network.neutron [-] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 814.062223] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 814.062883] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 814.063467] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 814.074177] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1758721f-343c-40fa-b05d-e55ee1415e47 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.081050] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452d31e6-c062-42a4-a40d-c8d012f16994 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.099841] env[65522]: DEBUG nova.compute.provider_tree [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.127898] env[65522]: WARNING openstack [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 814.129023] env[65522]: WARNING openstack [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 814.152487] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 814.184446] env[65522]: DEBUG nova.network.neutron [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 814.195334] env[65522]: DEBUG oslo_vmware.api [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114036, 'name': PowerOnVM_Task, 'duration_secs': 1.336878} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.195982] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 814.196285] env[65522]: INFO nova.compute.manager [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Took 6.47 seconds to spawn the instance on the hypervisor. [ 814.196803] env[65522]: DEBUG nova.compute.manager [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 814.197775] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3763d5-354d-4881-b57f-92fcfac799a9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.200949] env[65522]: INFO nova.compute.manager [-] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Took 1.38 seconds to deallocate network for instance. [ 814.213572] env[65522]: WARNING openstack [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 814.213950] env[65522]: WARNING openstack [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 814.322174] env[65522]: WARNING neutronclient.v2_0.client [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 814.322847] env[65522]: WARNING openstack [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 814.323283] env[65522]: WARNING openstack [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 814.400431] env[65522]: DEBUG oslo_vmware.api [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Task: {'id': task-5114042, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274106} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.400958] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 814.401360] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 814.401360] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 814.401510] env[65522]: INFO nova.compute.manager [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Took 1.17 seconds to destroy the instance on the hypervisor. [ 814.401832] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 814.402080] env[65522]: DEBUG nova.compute.manager [-] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 814.402168] env[65522]: DEBUG nova.network.neutron [-] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 814.402430] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 814.403201] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 814.403459] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 814.465340] env[65522]: DEBUG nova.network.neutron [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Updating instance_info_cache with network_info: [{"id": "c7aacbb2-e620-48af-838a-6103f858f4c8", "address": "fa:16:3e:ef:bc:77", "network": {"id": "d0ade77e-db62-42ae-9058-c33fcab1d2f9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1207104058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "234a226c969b4965bfa2c411bb4c8d90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7aacbb2-e6", "ovs_interfaceid": "c7aacbb2-e620-48af-838a-6103f858f4c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 814.471160] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 814.605555] env[65522]: DEBUG nova.scheduler.client.report [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 814.659694] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquiring lock "9fa6e208-f913-4ecc-8548-ed6503fe6273" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.660045] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Lock "9fa6e208-f913-4ecc-8548-ed6503fe6273" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.660289] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquiring lock "9fa6e208-f913-4ecc-8548-ed6503fe6273-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.660519] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Lock "9fa6e208-f913-4ecc-8548-ed6503fe6273-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.660750] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Lock "9fa6e208-f913-4ecc-8548-ed6503fe6273-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 814.664713] env[65522]: INFO nova.compute.manager [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Terminating instance [ 814.724044] env[65522]: DEBUG oslo_concurrency.lockutils [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.730495] env[65522]: INFO nova.compute.manager [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Took 38.22 seconds to build instance. [ 814.901750] env[65522]: DEBUG nova.network.neutron [-] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 814.971031] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Releasing lock "refresh_cache-05319f69-2ac7-4649-a90a-3542d87a3e82" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 814.971031] env[65522]: DEBUG nova.compute.manager [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Instance network_info: |[{"id": "c7aacbb2-e620-48af-838a-6103f858f4c8", "address": "fa:16:3e:ef:bc:77", "network": {"id": "d0ade77e-db62-42ae-9058-c33fcab1d2f9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1207104058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "234a226c969b4965bfa2c411bb4c8d90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7aacbb2-e6", "ovs_interfaceid": "c7aacbb2-e620-48af-838a-6103f858f4c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 814.971031] env[65522]: DEBUG oslo_concurrency.lockutils [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] Acquired lock "refresh_cache-05319f69-2ac7-4649-a90a-3542d87a3e82" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 814.971031] env[65522]: DEBUG nova.network.neutron [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Refreshing network info cache for port c7aacbb2-e620-48af-838a-6103f858f4c8 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 814.971031] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:bc:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '29945b49-d2b8-449d-9531-437917f49839', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7aacbb2-e620-48af-838a-6103f858f4c8', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 814.979914] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 814.981383] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 814.981816] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-423ea724-891b-4d45-8b2d-8591c980b42b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.006546] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 815.006546] env[65522]: value = "task-5114043" [ 815.006546] env[65522]: _type = "Task" [ 815.006546] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.018447] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114043, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.113811] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.355s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 815.116404] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.803s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 815.116685] env[65522]: DEBUG nova.objects.instance [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lazy-loading 'resources' on Instance uuid 4238d766-2a5d-40d7-b2f6-5cb9a5354760 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 815.137103] env[65522]: INFO nova.scheduler.client.report [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Deleted allocations for instance ea5b8e29-ebe6-444b-bd2d-029201003ee6 [ 815.170810] env[65522]: DEBUG nova.compute.manager [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 815.171074] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 815.172211] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e542e460-7a91-4bd1-8b9c-6ee0c4b58acd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.175520] env[65522]: DEBUG nova.network.neutron [-] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 815.184956] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 815.185084] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5caa78d-198d-4e2d-aed3-e53bc36acfbb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.194891] env[65522]: DEBUG oslo_vmware.api [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Waiting for the task: (returnval){ [ 815.194891] env[65522]: value = "task-5114044" [ 815.194891] env[65522]: _type = "Task" [ 815.194891] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.205311] env[65522]: DEBUG oslo_vmware.api [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5114044, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.233332] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d6dbc9be-82c6-4e95-965b-5bfa895016df tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Lock "1be995d0-3141-4dcd-ad9a-278d9874e457" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.739s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 815.319505] env[65522]: DEBUG nova.compute.manager [None req-a55ef713-46c6-46e8-8f2b-8efc4771da39 tempest-ServerDiagnosticsV248Test-661875750 tempest-ServerDiagnosticsV248Test-661875750-project-admin] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 815.320870] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d49960d5-662e-42b7-9e9a-25bbabda0dce {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.330211] env[65522]: INFO nova.compute.manager [None req-a55ef713-46c6-46e8-8f2b-8efc4771da39 tempest-ServerDiagnosticsV248Test-661875750 tempest-ServerDiagnosticsV248Test-661875750-project-admin] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Retrieving diagnostics [ 815.331140] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c4029e-2d40-4679-875f-a2bbafe3d535 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.404451] env[65522]: INFO nova.compute.manager [-] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Took 1.34 seconds to deallocate network for instance. [ 815.481983] env[65522]: WARNING neutronclient.v2_0.client [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 815.482813] env[65522]: WARNING openstack [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 815.483285] env[65522]: WARNING openstack [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 815.521384] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114043, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.616245] env[65522]: WARNING openstack [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 815.616639] env[65522]: WARNING openstack [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 815.632193] env[65522]: DEBUG nova.compute.manager [req-7c97a409-4cf7-4245-beff-0a4dd9c0d312 req-c347d58d-b560-4889-bf1b-672b42ff0c9c service nova] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Received event network-vif-deleted-012f75e3-8bea-4de0-9d56-d0b81de150d3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 815.632387] env[65522]: DEBUG nova.compute.manager [req-7c97a409-4cf7-4245-beff-0a4dd9c0d312 req-c347d58d-b560-4889-bf1b-672b42ff0c9c service nova] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Received event network-vif-deleted-ad290ac3-4edd-4dd7-b683-0f0283c2dcc4 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 815.650731] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3910a2e-3e4f-4b85-9307-c198fae59d5b tempest-ServerDiagnosticsTest-612509888 tempest-ServerDiagnosticsTest-612509888-project-member] Lock "ea5b8e29-ebe6-444b-bd2d-029201003ee6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.886s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 815.679240] env[65522]: INFO nova.compute.manager [-] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Took 1.28 seconds to deallocate network for instance. [ 815.687678] env[65522]: WARNING neutronclient.v2_0.client [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 815.688422] env[65522]: WARNING openstack [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 815.689431] env[65522]: WARNING openstack [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 815.711320] env[65522]: DEBUG oslo_vmware.api [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5114044, 'name': PowerOffVM_Task, 'duration_secs': 0.324119} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.714196] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 815.714379] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 815.715512] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-42de7c5b-8831-4d75-b6f0-bcf70ca29430 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.789837] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 815.790170] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 815.790765] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Deleting the datastore file [datastore2] 9fa6e208-f913-4ecc-8548-ed6503fe6273 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 815.790765] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-93b199d9-9cc7-4f09-bde0-43a530f8411c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.796214] env[65522]: DEBUG nova.network.neutron [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Updated VIF entry in instance network info cache for port c7aacbb2-e620-48af-838a-6103f858f4c8. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 815.796759] env[65522]: DEBUG nova.network.neutron [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Updating instance_info_cache with network_info: [{"id": "c7aacbb2-e620-48af-838a-6103f858f4c8", "address": "fa:16:3e:ef:bc:77", "network": {"id": "d0ade77e-db62-42ae-9058-c33fcab1d2f9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1207104058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "234a226c969b4965bfa2c411bb4c8d90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7aacbb2-e6", "ovs_interfaceid": "c7aacbb2-e620-48af-838a-6103f858f4c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 815.800861] env[65522]: DEBUG oslo_vmware.api [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Waiting for the task: (returnval){ [ 815.800861] env[65522]: value = "task-5114046" [ 815.800861] env[65522]: _type = "Task" [ 815.800861] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.814868] env[65522]: DEBUG oslo_vmware.api [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5114046, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.915118] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 815.997621] env[65522]: DEBUG nova.compute.manager [req-08767613-b0b4-44a7-84a3-16b065ee8a81 req-0837aa5e-9cea-4e10-9e4c-35c1da3de741 service nova] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Received event network-vif-deleted-4c16f1be-de7b-45f4-9510-b1e2d358c37c {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 816.026202] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114043, 'name': CreateVM_Task, 'duration_secs': 0.56805} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.026659] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 816.027102] env[65522]: WARNING neutronclient.v2_0.client [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 816.027557] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.027721] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 816.028098] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 816.028409] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33d6cd9d-b9dc-4b2c-b3a7-8bcae7db7f33 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.037252] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 816.037252] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52af947a-0b76-dd00-09fe-4ff9024afc21" [ 816.037252] env[65522]: _type = "Task" [ 816.037252] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.049923] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52af947a-0b76-dd00-09fe-4ff9024afc21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.144574] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529936a3-c101-b946-9e53-8bfd69a815a2/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 816.145771] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f785204-d60f-4115-81cb-40f6c9a1fe3a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.152574] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529936a3-c101-b946-9e53-8bfd69a815a2/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 816.152765] env[65522]: ERROR oslo_vmware.rw_handles [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529936a3-c101-b946-9e53-8bfd69a815a2/disk-0.vmdk due to incomplete transfer. [ 816.155700] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6c1acd4f-2962-4a9f-b3a7-fc14c77bfd3f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.167965] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529936a3-c101-b946-9e53-8bfd69a815a2/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 816.167965] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Uploaded image 987d36cc-3327-435f-9eb6-6f994a08ebfa to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 816.169833] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 816.172465] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-55d7a81b-9e7c-4bde-9685-97fa8b6ed48a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.180276] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 816.180276] env[65522]: value = "task-5114047" [ 816.180276] env[65522]: _type = "Task" [ 816.180276] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.189955] env[65522]: DEBUG oslo_concurrency.lockutils [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.190300] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114047, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.233944] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14fb5380-5d67-4f9f-97a3-6f171e00bf61 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.245128] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e1fdcf-fad3-4692-8f9e-85f4b1802d56 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.294082] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827701c2-3eeb-4f39-ac05-bc4aeffc031d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.302610] env[65522]: DEBUG oslo_concurrency.lockutils [req-cfd9daa1-efef-448a-98c3-a6026761e010 req-8e99369d-054e-4be0-b986-e340e455ed3f service nova] Releasing lock "refresh_cache-05319f69-2ac7-4649-a90a-3542d87a3e82" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 816.304166] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d429606d-28ef-47d4-be37-e59e85f02b5f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.326027] env[65522]: DEBUG nova.compute.provider_tree [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.331620] env[65522]: DEBUG oslo_vmware.api [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Task: {'id': task-5114046, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.267487} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.332012] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 816.332369] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 816.332498] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.332620] env[65522]: INFO nova.compute.manager [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Took 1.16 seconds to destroy the instance on the hypervisor. [ 816.332903] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 816.333451] env[65522]: DEBUG nova.compute.manager [-] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 816.333551] env[65522]: DEBUG nova.network.neutron [-] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 816.333839] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 816.336688] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.336688] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.448738] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 816.551374] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52af947a-0b76-dd00-09fe-4ff9024afc21, 'name': SearchDatastore_Task, 'duration_secs': 0.016584} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.551875] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 816.552878] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 816.552878] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.552878] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 816.553178] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 816.553346] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0289d642-8dcb-4ff4-ac25-17f6aaa62933 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.563182] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 816.563437] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 816.564325] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91d5648f-9b70-4723-b454-761d61dfd05a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.571057] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 816.571057] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d919d9-1cb5-b69f-927e-f8f060d59636" [ 816.571057] env[65522]: _type = "Task" [ 816.571057] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.583358] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d919d9-1cb5-b69f-927e-f8f060d59636, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.692601] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114047, 'name': Destroy_Task, 'duration_secs': 0.415906} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.692805] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Destroyed the VM [ 816.693838] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 816.693838] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c23883de-750b-4974-a6e1-6b5064afef89 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.702054] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 816.702054] env[65522]: value = "task-5114048" [ 816.702054] env[65522]: _type = "Task" [ 816.702054] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.714905] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114048, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.834253] env[65522]: DEBUG nova.scheduler.client.report [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 817.082732] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d919d9-1cb5-b69f-927e-f8f060d59636, 'name': SearchDatastore_Task, 'duration_secs': 0.011252} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.083606] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b00aeec3-fa6f-40d4-967d-1c4f4cc29447 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.090846] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 817.090846] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520b4a8a-24f2-0783-3895-5bef308f033e" [ 817.090846] env[65522]: _type = "Task" [ 817.090846] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.101903] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520b4a8a-24f2-0783-3895-5bef308f033e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.177402] env[65522]: DEBUG nova.network.neutron [-] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 817.212496] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114048, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.340713] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.224s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 817.343294] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.460s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 817.344149] env[65522]: DEBUG nova.objects.instance [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Lazy-loading 'resources' on Instance uuid fdf56fa6-ad18-499e-a171-6242abb293bf {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 817.368066] env[65522]: INFO nova.scheduler.client.report [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Deleted allocations for instance 4238d766-2a5d-40d7-b2f6-5cb9a5354760 [ 817.602107] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520b4a8a-24f2-0783-3895-5bef308f033e, 'name': SearchDatastore_Task, 'duration_secs': 0.018114} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.602417] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 817.602683] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 05319f69-2ac7-4649-a90a-3542d87a3e82/05319f69-2ac7-4649-a90a-3542d87a3e82.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 817.602959] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89762995-080c-49dd-b99b-93fa34afab56 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.611068] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 817.611068] env[65522]: value = "task-5114049" [ 817.611068] env[65522]: _type = "Task" [ 817.611068] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.620938] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114049, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.683513] env[65522]: INFO nova.compute.manager [-] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Took 1.35 seconds to deallocate network for instance. [ 817.714442] env[65522]: DEBUG oslo_vmware.api [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114048, 'name': RemoveSnapshot_Task, 'duration_secs': 0.554357} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.714756] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 817.715025] env[65522]: INFO nova.compute.manager [None req-d011cc2b-b1ee-4179-81a3-534af83ec29e tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Took 17.42 seconds to snapshot the instance on the hypervisor. [ 817.793491] env[65522]: DEBUG nova.compute.manager [req-eebce1e4-08ee-4b1c-aa1c-51386d94ee55 req-8e4a1c61-458f-4b79-a59d-1b2c36878e7d service nova] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Received event network-vif-deleted-7e072218-2d88-437a-b6fb-597ced883532 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 817.880044] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8828fc12-0bf8-4a06-92a0-34272e1c480e tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "4238d766-2a5d-40d7-b2f6-5cb9a5354760" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.326s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 818.122772] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114049, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.195688] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 818.392456] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8605f500-5a93-4222-a12d-68550672986b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.400606] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0abafe17-d146-4b98-b520-e2f75bd73a56 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.435092] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f324104c-abee-4086-a911-70dd5e54a31a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.446325] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bbb280d-ec50-4d78-8676-f84fd924a0ab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.466342] env[65522]: DEBUG nova.compute.provider_tree [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.508194] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Acquiring lock "8eda6816-6781-4ad9-9210-11fa623f24b0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 818.508787] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Lock "8eda6816-6781-4ad9-9210-11fa623f24b0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 818.622564] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114049, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.970521] env[65522]: DEBUG nova.scheduler.client.report [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 819.011274] env[65522]: DEBUG nova.compute.manager [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 819.124420] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114049, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.475656] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.132s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 819.478021] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 33.017s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 819.510231] env[65522]: INFO nova.scheduler.client.report [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Deleted allocations for instance fdf56fa6-ad18-499e-a171-6242abb293bf [ 819.529561] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 819.624271] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114049, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.551476} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.624541] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 05319f69-2ac7-4649-a90a-3542d87a3e82/05319f69-2ac7-4649-a90a-3542d87a3e82.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 819.624752] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 819.625018] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-73dd28cd-f064-4663-8454-c82a79d47f46 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.633080] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 819.633080] env[65522]: value = "task-5114050" [ 819.633080] env[65522]: _type = "Task" [ 819.633080] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.645174] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114050, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.020110] env[65522]: DEBUG oslo_concurrency.lockutils [None req-50fa17a0-d270-49de-a7e8-57bf74e3ad5c tempest-ServersTestJSON-779093636 tempest-ServersTestJSON-779093636-project-member] Lock "fdf56fa6-ad18-499e-a171-6242abb293bf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.913s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 820.143266] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114050, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069131} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.143574] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 820.144374] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2295f3ad-efeb-4166-9da4-b6468916eba7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.167466] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 05319f69-2ac7-4649-a90a-3542d87a3e82/05319f69-2ac7-4649-a90a-3542d87a3e82.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.167778] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6890dc5-b53e-4c09-8494-cc1077bb0986 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.187835] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 820.187835] env[65522]: value = "task-5114051" [ 820.187835] env[65522]: _type = "Task" [ 820.187835] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.197164] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114051, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.325336] env[65522]: DEBUG nova.compute.manager [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 820.326335] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9e7335-5b17-4e05-9d75-113bad46b25e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.498021] env[65522]: INFO nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating resource usage from migration 98e33664-f395-480b-a10d-3904a0963458 [ 820.522219] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance b2d70982-54e7-459c-a0d8-48bf4b6e4345 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.522419] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7c6f8218-602d-44f3-8012-de5a96972785 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.522542] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.522688] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance e5d2cc50-4ee6-4e64-9b52-888968a717ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.522794] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance a8f9f223-023a-48fc-9186-a360a78e27a9 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.522916] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 97dc8cac-8241-4912-a3ed-689439ef7ff8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.523028] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance ca964440-5375-4aff-8b45-96fbe829dd16 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.523157] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance cd82f2fd-d657-4ece-8e90-9a47072f5546 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.523266] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 87a3c63d-794d-44ab-bad6-65c323d72ae7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.523377] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 68d8d59b-4fc8-4993-8c9e-31b2329c88dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.523558] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 75c8848b-aea9-43f1-8697-9224050d1fef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.523626] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 42a08cfb-d865-4967-a086-370a2ca98b7d actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.523732] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 361e54c7-4309-48b3-ad3e-b2ebc783741c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.523853] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.523971] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 028c7fc5-5624-4dd5-9e2f-48191a86f765 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.524105] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance d5e8dd05-dc3c-4831-b4b0-ac100360f3e7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.524198] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance da211708-a6a2-4e03-bb41-c5b93564de20 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.524319] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 1e8088ef-6354-40a0-95ee-06732e7f088f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.524429] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance bb20d99f-fc49-47bf-a6b7-74ea966b71eb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.524538] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 24f520ec-6a56-4f17-9ae4-2c856bf99582 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.524657] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7693f017-69b3-4ec4-ae3f-7e65ed34a471 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.524772] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance d2fdffb8-76d1-4b74-819f-25eb29ca91ce is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.524895] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 1eebb320-30e2-4d8b-b1fd-19e31fb35b77 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.524992] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance cdee3a32-ad51-4364-8440-4744e8df322f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.525138] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 9fa6e208-f913-4ecc-8548-ed6503fe6273 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 820.525250] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 1be995d0-3141-4dcd-ad9a-278d9874e457 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.525356] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 05319f69-2ac7-4649-a90a-3542d87a3e82 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 820.698415] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114051, 'name': ReconfigVM_Task, 'duration_secs': 0.286348} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.698710] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 05319f69-2ac7-4649-a90a-3542d87a3e82/05319f69-2ac7-4649-a90a-3542d87a3e82.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.699039] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Creating Virtual Disk of size 1048576 KB and adapter type paraVirtual on the data store {{(pid=65522) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1383}} [ 820.699729] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CreateVirtualDisk_Task with opID=oslo.vmware-202cfbf3-bc5c-4dd0-9b54-35b713d25158 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.707901] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 820.707901] env[65522]: value = "task-5114052" [ 820.707901] env[65522]: _type = "Task" [ 820.707901] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.718428] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114052, 'name': CreateVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.839431] env[65522]: INFO nova.compute.manager [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] instance snapshotting [ 820.842982] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2c347d-c5ec-43c3-a8e4-4c8ff794f2be {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.865552] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0a4461-bc39-4aa8-9741-35669e8a9730 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.028388] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 18568baa-83de-4b54-addf-dee66211c8d7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 821.218948] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114052, 'name': CreateVirtualDisk_Task, 'duration_secs': 0.054473} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.219413] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Created Virtual Disk of size 1048576 KB and type thin {{(pid=65522) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1404}} [ 821.220343] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b1f2aa4-6817-4354-a2f9-0ff4dcdc7c37 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.249237] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 05319f69-2ac7-4649-a90a-3542d87a3e82/ephemeral_0.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.249568] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b832f57a-258b-4f75-9639-de578e6a9a76 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.269155] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 821.269155] env[65522]: value = "task-5114053" [ 821.269155] env[65522]: _type = "Task" [ 821.269155] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.278164] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114053, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.383301] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 821.383705] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3395cd78-00ba-4e65-9797-e0548568d800 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.394100] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 821.394100] env[65522]: value = "task-5114054" [ 821.394100] env[65522]: _type = "Task" [ 821.394100] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.404634] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114054, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.532080] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 0b48b747-149c-4163-bdd0-7d0b07ddacfd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 821.781519] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114053, 'name': ReconfigVM_Task, 'duration_secs': 0.308377} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.781850] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 05319f69-2ac7-4649-a90a-3542d87a3e82/ephemeral_0.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.782538] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-63df492d-4560-4b9c-b051-68ad75e2df6e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.793467] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 821.793467] env[65522]: value = "task-5114055" [ 821.793467] env[65522]: _type = "Task" [ 821.793467] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.803403] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114055, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.904324] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114054, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.036576] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 880390f5-00f1-4dea-9345-29c06289b340 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 822.304405] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114055, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.406526] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114054, 'name': CreateSnapshot_Task, 'duration_secs': 0.578638} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.406811] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 822.407959] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7060e641-d4c6-4f04-bbd4-666df0fbdb6a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.540735] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 77b3ca48-d2c6-4743-a44a-53b9aa84662a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 822.808557] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114055, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.929059] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 822.929198] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6fddc8b0-54ef-4f30-bc41-82f6dfe55f69 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.939753] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 822.939753] env[65522]: value = "task-5114056" [ 822.939753] env[65522]: _type = "Task" [ 822.939753] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.949614] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114056, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.044702] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7617189c-a902-42e7-8165-0e7c4a1de06d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 823.044976] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Migration 98e33664-f395-480b-a10d-3904a0963458 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 823.045106] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 495fa98f-a9db-4214-87cc-d29209d3cb62 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 823.306101] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114055, 'name': Rename_Task, 'duration_secs': 1.185487} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.306101] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 823.306101] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-892890ba-083b-497b-a315-014ac35d8f8c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.314469] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 823.314469] env[65522]: value = "task-5114057" [ 823.314469] env[65522]: _type = "Task" [ 823.314469] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.324632] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114057, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.451992] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114056, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.548527] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 26432ad6-610e-49c2-8699-f1e7e128e5b8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 823.826304] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114057, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.952065] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114056, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.053264] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 5e227f6c-9f13-416f-8e6d-2f7d931619fd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 824.325264] env[65522]: DEBUG oslo_vmware.api [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114057, 'name': PowerOnVM_Task, 'duration_secs': 0.535219} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.325701] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 824.325794] env[65522]: INFO nova.compute.manager [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Took 11.37 seconds to spawn the instance on the hypervisor. [ 824.325974] env[65522]: DEBUG nova.compute.manager [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 824.326819] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05629403-b3a2-43bc-8086-a8d4d478e382 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.453056] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114056, 'name': CloneVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.556796] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 8eda6816-6781-4ad9-9210-11fa623f24b0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 824.557114] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 18 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 824.557290] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4032MB phys_disk=100GB used_disk=18GB total_vcpus=48 used_vcpus=18 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '18', 'num_vm_active': '16', 'num_task_None': '15', 'num_os_type_None': '18', 'num_proj_fa11b46d9fe144f391233e6eb9c819d7': '1', 'io_workload': '2', 'num_proj_7340f053e9784f30be39030d99fe698f': '1', 'num_proj_95bca233019f4e2f8bac0dd800392a69': '2', 'num_proj_9b8112ba49034e6fabcb3fbbd46edf41': '3', 'num_vm_rescued': '1', 'num_proj_eb9a9e5da1194ab7927f72d21d1484b0': '1', 'num_proj_503adb79a1f74f1b98e1f06ec54f1584': '1', 'num_proj_37fc01876567476f9b93d765b2cfddc8': '1', 'num_proj_6d64b94285334dd6b26928cabedbd67c': '1', 'num_task_image_snapshot_pending': '1', 'num_proj_bfbeddbba72d447d93831a4b381f0ea0': '2', 'num_proj_b5c14a856d454048856539895f220c2c': '2', 'num_task_resize_prep': '1', 'num_proj_6ecccb656b0d4c96b40b200cdcddbad5': '1', 'num_proj_3ae929571fb5465cba5bedffbb6aa4c8': '1', 'num_vm_building': '1', 'num_task_spawning': '1', 'num_proj_234a226c969b4965bfa2c411bb4c8d90': '1'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 824.850488] env[65522]: INFO nova.compute.manager [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Took 43.65 seconds to build instance. [ 824.956759] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114056, 'name': CloneVM_Task, 'duration_secs': 1.518425} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.957231] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Created linked-clone VM from snapshot [ 824.958153] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-199aec76-0ea2-436f-b579-716651101485 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.969561] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Uploading image eb493893-0d43-4227-a4d0-f0c68d713fef {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 824.996587] env[65522]: DEBUG oslo_vmware.rw_handles [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 824.996587] env[65522]: value = "vm-994799" [ 824.996587] env[65522]: _type = "VirtualMachine" [ 824.996587] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 824.996883] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-868c0fa8-cf47-4b56-a827-22aa6774a715 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.005690] env[65522]: DEBUG oslo_vmware.rw_handles [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lease: (returnval){ [ 825.005690] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5241055a-e34b-02be-186a-2c0a51d220ba" [ 825.005690] env[65522]: _type = "HttpNfcLease" [ 825.005690] env[65522]: } obtained for exporting VM: (result){ [ 825.005690] env[65522]: value = "vm-994799" [ 825.005690] env[65522]: _type = "VirtualMachine" [ 825.005690] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 825.005931] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the lease: (returnval){ [ 825.005931] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5241055a-e34b-02be-186a-2c0a51d220ba" [ 825.005931] env[65522]: _type = "HttpNfcLease" [ 825.005931] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 825.014330] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 825.014330] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5241055a-e34b-02be-186a-2c0a51d220ba" [ 825.014330] env[65522]: _type = "HttpNfcLease" [ 825.014330] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 825.075275] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165e7b75-058c-4881-a99c-9829b0cd106b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.083287] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f017f6b-69b2-4976-ab16-d415cbcd13ad {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.113834] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76025f48-40b2-48e9-b6e8-32b10d6ceabb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.122226] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eeaedc0-0a91-42c2-827c-92079f5e8555 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.136176] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.354373] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef752d20-e941-46c7-ab2f-2e8f29ef846f tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "05319f69-2ac7-4649-a90a-3542d87a3e82" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.176s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 825.514871] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 825.514871] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5241055a-e34b-02be-186a-2c0a51d220ba" [ 825.514871] env[65522]: _type = "HttpNfcLease" [ 825.514871] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 825.515189] env[65522]: DEBUG oslo_vmware.rw_handles [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 825.515189] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5241055a-e34b-02be-186a-2c0a51d220ba" [ 825.515189] env[65522]: _type = "HttpNfcLease" [ 825.515189] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 825.515948] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcdb1783-3d0c-4c78-b026-828dd24d0cb6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.524337] env[65522]: DEBUG oslo_vmware.rw_handles [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d401b6-91b6-02f5-05dc-d7ec890d3309/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 825.524599] env[65522]: DEBUG oslo_vmware.rw_handles [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d401b6-91b6-02f5-05dc-d7ec890d3309/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 825.640118] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 825.644235] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-26ff74b3-d19c-4259-a040-8da5a20909d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.880961] env[65522]: DEBUG nova.compute.manager [None req-e013b9c1-e032-4dd1-9f5a-ba2f657775cc tempest-ServerDiagnosticsV248Test-661875750 tempest-ServerDiagnosticsV248Test-661875750-project-admin] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 825.884605] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-626eb500-9970-4271-8bf3-0782d3424a55 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.893438] env[65522]: INFO nova.compute.manager [None req-e013b9c1-e032-4dd1-9f5a-ba2f657775cc tempest-ServerDiagnosticsV248Test-661875750 tempest-ServerDiagnosticsV248Test-661875750-project-admin] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Retrieving diagnostics [ 825.894910] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985c02e3-618d-49e7-b8ce-99a0bb4f33f8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.148138] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 826.148542] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.671s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 826.148885] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.581s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 826.151292] env[65522]: INFO nova.compute.claims [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.381598] env[65522]: DEBUG nova.compute.manager [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Received event network-changed-c7aacbb2-e620-48af-838a-6103f858f4c8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 826.383182] env[65522]: DEBUG nova.compute.manager [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Refreshing instance network info cache due to event network-changed-c7aacbb2-e620-48af-838a-6103f858f4c8. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 826.383182] env[65522]: DEBUG oslo_concurrency.lockutils [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] Acquiring lock "refresh_cache-05319f69-2ac7-4649-a90a-3542d87a3e82" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.383182] env[65522]: DEBUG oslo_concurrency.lockutils [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] Acquired lock "refresh_cache-05319f69-2ac7-4649-a90a-3542d87a3e82" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 826.383182] env[65522]: DEBUG nova.network.neutron [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Refreshing network info cache for port c7aacbb2-e620-48af-838a-6103f858f4c8 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 826.796158] env[65522]: DEBUG oslo_concurrency.lockutils [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Acquiring lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 826.797077] env[65522]: DEBUG oslo_concurrency.lockutils [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 826.887036] env[65522]: WARNING neutronclient.v2_0.client [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 826.887864] env[65522]: WARNING openstack [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 826.888359] env[65522]: WARNING openstack [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 827.009148] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Acquiring lock "1be995d0-3141-4dcd-ad9a-278d9874e457" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 827.009518] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Lock "1be995d0-3141-4dcd-ad9a-278d9874e457" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 827.009925] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Acquiring lock "1be995d0-3141-4dcd-ad9a-278d9874e457-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 827.010194] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Lock "1be995d0-3141-4dcd-ad9a-278d9874e457-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 827.010500] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Lock "1be995d0-3141-4dcd-ad9a-278d9874e457-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 827.012979] env[65522]: INFO nova.compute.manager [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Terminating instance [ 827.033226] env[65522]: WARNING openstack [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 827.033487] env[65522]: WARNING openstack [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 827.104301] env[65522]: WARNING neutronclient.v2_0.client [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 827.105178] env[65522]: WARNING openstack [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 827.105673] env[65522]: WARNING openstack [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 827.194998] env[65522]: DEBUG nova.network.neutron [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Updated VIF entry in instance network info cache for port c7aacbb2-e620-48af-838a-6103f858f4c8. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 827.195527] env[65522]: DEBUG nova.network.neutron [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Updating instance_info_cache with network_info: [{"id": "c7aacbb2-e620-48af-838a-6103f858f4c8", "address": "fa:16:3e:ef:bc:77", "network": {"id": "d0ade77e-db62-42ae-9058-c33fcab1d2f9", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-1207104058-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "234a226c969b4965bfa2c411bb4c8d90", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7aacbb2-e6", "ovs_interfaceid": "c7aacbb2-e620-48af-838a-6103f858f4c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 827.301451] env[65522]: DEBUG nova.compute.utils [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 827.518844] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Acquiring lock "refresh_cache-1be995d0-3141-4dcd-ad9a-278d9874e457" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.519213] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Acquired lock "refresh_cache-1be995d0-3141-4dcd-ad9a-278d9874e457" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 827.519316] env[65522]: DEBUG nova.network.neutron [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 827.664294] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df668979-4609-4821-ba2f-49a8c5a2299a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.673489] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe57f85-10cc-48bc-8cc3-2e5f655ed5eb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.705622] env[65522]: DEBUG oslo_concurrency.lockutils [req-61a8e58a-b5c5-4dc5-8d0c-9866e3708465 req-0925f9fe-8ca7-49c0-ac1c-6352aa71a260 service nova] Releasing lock "refresh_cache-05319f69-2ac7-4649-a90a-3542d87a3e82" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 827.706928] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2e2476-805b-41a1-800f-29b326863fdf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.716796] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de873e44-1c8f-4a46-b890-69d4d8045f78 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.735481] env[65522]: DEBUG nova.compute.provider_tree [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.805013] env[65522]: DEBUG oslo_concurrency.lockutils [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 828.022427] env[65522]: WARNING neutronclient.v2_0.client [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 828.023169] env[65522]: WARNING openstack [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.023555] env[65522]: WARNING openstack [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.047772] env[65522]: DEBUG nova.network.neutron [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 828.110515] env[65522]: DEBUG nova.network.neutron [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 828.238944] env[65522]: DEBUG nova.scheduler.client.report [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 828.617328] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Releasing lock "refresh_cache-1be995d0-3141-4dcd-ad9a-278d9874e457" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 828.617328] env[65522]: DEBUG nova.compute.manager [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 828.617328] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 828.617328] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1daa5441-3bad-4332-b54d-5b6574822567 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.626629] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 828.626629] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-663809fb-497f-4bd1-b01a-19a6f51adf77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.633761] env[65522]: DEBUG oslo_vmware.api [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Waiting for the task: (returnval){ [ 828.633761] env[65522]: value = "task-5114059" [ 828.633761] env[65522]: _type = "Task" [ 828.633761] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.643101] env[65522]: DEBUG oslo_vmware.api [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114059, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.697266] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "7c6f8218-602d-44f3-8012-de5a96972785" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 828.697545] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "7c6f8218-602d-44f3-8012-de5a96972785" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 828.697794] env[65522]: INFO nova.compute.manager [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Rebooting instance [ 828.745015] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.596s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 828.745260] env[65522]: DEBUG nova.compute.manager [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 828.749356] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.520s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 828.750009] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 828.753093] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.640s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 828.754574] env[65522]: INFO nova.compute.claims [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.781849] env[65522]: INFO nova.scheduler.client.report [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Deleted allocations for instance 1e8088ef-6354-40a0-95ee-06732e7f088f [ 828.879607] env[65522]: DEBUG oslo_concurrency.lockutils [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Acquiring lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 828.880163] env[65522]: DEBUG oslo_concurrency.lockutils [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 828.880631] env[65522]: INFO nova.compute.manager [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Attaching volume ec224e94-bcce-4815-96c0-d6f0d022ccbb to /dev/sdb [ 828.925351] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8edaaa6d-1d42-4c93-86e3-ea4e5f9c40a1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.933661] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8badfa6d-8e8f-4639-b85f-6360dc598bcc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.950112] env[65522]: DEBUG nova.virt.block_device [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Updating existing volume attachment record: 097ff7fd-1893-4a06-9486-2c05d7be4d2a {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 829.146158] env[65522]: DEBUG oslo_vmware.api [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114059, 'name': PowerOffVM_Task, 'duration_secs': 0.149568} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.146158] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 829.146158] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 829.146377] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6037c639-7426-40f4-8b30-9e9d4f550d2c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.173747] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 829.174052] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 829.174254] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Deleting the datastore file [datastore1] 1be995d0-3141-4dcd-ad9a-278d9874e457 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 829.174533] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-56519154-ccbe-4b34-a85f-6a72f9ef2b21 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.182714] env[65522]: DEBUG oslo_vmware.api [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Waiting for the task: (returnval){ [ 829.182714] env[65522]: value = "task-5114064" [ 829.182714] env[65522]: _type = "Task" [ 829.182714] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.192317] env[65522]: DEBUG oslo_vmware.api [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114064, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.215588] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.215848] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 829.216091] env[65522]: DEBUG nova.network.neutron [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 829.263366] env[65522]: DEBUG nova.compute.utils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 829.265667] env[65522]: DEBUG nova.compute.manager [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 829.265868] env[65522]: DEBUG nova.network.neutron [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 829.266367] env[65522]: WARNING neutronclient.v2_0.client [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 829.266567] env[65522]: WARNING neutronclient.v2_0.client [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 829.267263] env[65522]: WARNING openstack [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.267694] env[65522]: WARNING openstack [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.292894] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f4aafbca-2f6a-4f70-9e1e-d2c6bcfd2253 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "1e8088ef-6354-40a0-95ee-06732e7f088f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 43.786s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 829.334745] env[65522]: DEBUG nova.policy [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7e3ca1e6d7145f49596c1ab6466ca01', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd6e7f7b6a224cac8b8e52bf19032911', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 829.671460] env[65522]: DEBUG nova.network.neutron [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Successfully created port: 990aae71-9e8e-4da6-9e57-5c9257d127a4 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 829.695867] env[65522]: DEBUG oslo_vmware.api [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Task: {'id': task-5114064, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169684} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.695867] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 829.696189] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 829.696265] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 829.696530] env[65522]: INFO nova.compute.manager [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Took 1.08 seconds to destroy the instance on the hypervisor. [ 829.696766] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 829.696972] env[65522]: DEBUG nova.compute.manager [-] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 829.697086] env[65522]: DEBUG nova.network.neutron [-] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 829.697336] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 829.697875] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.698139] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.718858] env[65522]: WARNING neutronclient.v2_0.client [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 829.719603] env[65522]: WARNING openstack [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.719968] env[65522]: WARNING openstack [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.727723] env[65522]: DEBUG nova.network.neutron [-] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 829.727939] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 829.766758] env[65522]: DEBUG nova.compute.manager [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 829.953685] env[65522]: WARNING openstack [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.954099] env[65522]: WARNING openstack [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.182983] env[65522]: WARNING neutronclient.v2_0.client [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 830.183669] env[65522]: WARNING openstack [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.184424] env[65522]: WARNING openstack [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.230202] env[65522]: DEBUG nova.network.neutron [-] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 830.327985] env[65522]: DEBUG nova.network.neutron [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance_info_cache with network_info: [{"id": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "address": "fa:16:3e:69:05:d7", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf16c93a-75", "ovs_interfaceid": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 830.337513] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17995be-91a8-4fee-984b-649473053e4e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.350020] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c658d1d2-b765-48a4-9aeb-d0a6e384c249 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.383076] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe4d2d7-7bde-46d3-9e3a-ae626c7d4c0a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.392709] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f4ba2e-5d58-4b16-8ee3-bbf4a760ff99 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.407335] env[65522]: DEBUG nova.compute.provider_tree [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.733102] env[65522]: INFO nova.compute.manager [-] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Took 1.04 seconds to deallocate network for instance. [ 830.787074] env[65522]: DEBUG nova.compute.manager [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 830.824634] env[65522]: DEBUG nova.virt.hardware [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 830.825288] env[65522]: DEBUG nova.virt.hardware [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 830.825712] env[65522]: DEBUG nova.virt.hardware [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 830.826202] env[65522]: DEBUG nova.virt.hardware [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 830.828011] env[65522]: DEBUG nova.virt.hardware [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 830.828011] env[65522]: DEBUG nova.virt.hardware [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 830.828011] env[65522]: DEBUG nova.virt.hardware [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.828205] env[65522]: DEBUG nova.virt.hardware [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 830.828362] env[65522]: DEBUG nova.virt.hardware [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 830.828710] env[65522]: DEBUG nova.virt.hardware [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 830.829125] env[65522]: DEBUG nova.virt.hardware [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 830.831883] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a64d368-b862-4fbd-9dc7-c48f5da2235c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.836623] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 830.846635] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e03d32-a8ad-4612-aa7e-2bc29603999c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.912082] env[65522]: DEBUG nova.scheduler.client.report [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 831.033879] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "179a37ed-1176-4082-8c68-1be6abd3fd9d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 831.034332] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "179a37ed-1176-4082-8c68-1be6abd3fd9d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 831.061879] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 831.062231] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 831.146917] env[65522]: DEBUG nova.compute.manager [req-84de6c72-e093-4c94-b351-3c8bc9004cd8 req-66b3e824-7089-4019-acf5-71e52883b659 service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Received event network-vif-plugged-990aae71-9e8e-4da6-9e57-5c9257d127a4 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 831.147243] env[65522]: DEBUG oslo_concurrency.lockutils [req-84de6c72-e093-4c94-b351-3c8bc9004cd8 req-66b3e824-7089-4019-acf5-71e52883b659 service nova] Acquiring lock "18568baa-83de-4b54-addf-dee66211c8d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 831.147483] env[65522]: DEBUG oslo_concurrency.lockutils [req-84de6c72-e093-4c94-b351-3c8bc9004cd8 req-66b3e824-7089-4019-acf5-71e52883b659 service nova] Lock "18568baa-83de-4b54-addf-dee66211c8d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 831.147672] env[65522]: DEBUG oslo_concurrency.lockutils [req-84de6c72-e093-4c94-b351-3c8bc9004cd8 req-66b3e824-7089-4019-acf5-71e52883b659 service nova] Lock "18568baa-83de-4b54-addf-dee66211c8d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 831.147868] env[65522]: DEBUG nova.compute.manager [req-84de6c72-e093-4c94-b351-3c8bc9004cd8 req-66b3e824-7089-4019-acf5-71e52883b659 service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] No waiting events found dispatching network-vif-plugged-990aae71-9e8e-4da6-9e57-5c9257d127a4 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 831.148062] env[65522]: WARNING nova.compute.manager [req-84de6c72-e093-4c94-b351-3c8bc9004cd8 req-66b3e824-7089-4019-acf5-71e52883b659 service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Received unexpected event network-vif-plugged-990aae71-9e8e-4da6-9e57-5c9257d127a4 for instance with vm_state building and task_state spawning. [ 831.237669] env[65522]: DEBUG nova.network.neutron [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Successfully updated port: 990aae71-9e8e-4da6-9e57-5c9257d127a4 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 831.242271] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 831.342122] env[65522]: DEBUG nova.compute.manager [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 831.343150] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0745bf10-41ee-4df4-99ee-fd053c8dee8e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.418115] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.665s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 831.418679] env[65522]: DEBUG nova.compute.manager [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 831.422360] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.790s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 831.424337] env[65522]: INFO nova.compute.claims [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 831.541025] env[65522]: DEBUG nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 831.564630] env[65522]: DEBUG nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 831.741548] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Acquiring lock "refresh_cache-18568baa-83de-4b54-addf-dee66211c8d7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.741895] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Acquired lock "refresh_cache-18568baa-83de-4b54-addf-dee66211c8d7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 831.741895] env[65522]: DEBUG nova.network.neutron [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 831.924104] env[65522]: DEBUG nova.compute.utils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 831.926943] env[65522]: DEBUG nova.compute.manager [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 831.927255] env[65522]: DEBUG nova.network.neutron [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 831.927862] env[65522]: WARNING neutronclient.v2_0.client [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 831.928269] env[65522]: WARNING neutronclient.v2_0.client [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 831.928904] env[65522]: WARNING openstack [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.929331] env[65522]: WARNING openstack [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.990053] env[65522]: DEBUG nova.policy [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2109e40451943a3884012e51bf118f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b5c14a856d454048856539895f220c2c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 832.062682] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 832.086374] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 832.250026] env[65522]: WARNING openstack [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 832.250522] env[65522]: WARNING openstack [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 832.288437] env[65522]: DEBUG nova.network.neutron [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 832.291754] env[65522]: DEBUG nova.network.neutron [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Successfully created port: 6f040560-31cf-4beb-916d-44d19ed60c9e {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 832.309469] env[65522]: WARNING openstack [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 832.309877] env[65522]: WARNING openstack [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 832.360661] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17bd569b-c6a4-4604-b90a-3932bad96736 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.368834] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Doing hard reboot of VM {{(pid=65522) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 832.369168] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-858d8fa2-de45-4fd4-b6db-3a5e061c25b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.378688] env[65522]: DEBUG oslo_vmware.api [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 832.378688] env[65522]: value = "task-5114066" [ 832.378688] env[65522]: _type = "Task" [ 832.378688] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.390377] env[65522]: DEBUG oslo_vmware.api [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114066, 'name': ResetVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.410209] env[65522]: WARNING neutronclient.v2_0.client [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 832.410923] env[65522]: WARNING openstack [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 832.411304] env[65522]: WARNING openstack [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 832.439064] env[65522]: DEBUG nova.compute.manager [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 832.532147] env[65522]: DEBUG nova.network.neutron [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Updating instance_info_cache with network_info: [{"id": "990aae71-9e8e-4da6-9e57-5c9257d127a4", "address": "fa:16:3e:1b:6a:ed", "network": {"id": "5165f03c-e790-4121-b5d7-3c5176afdd7e", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-551605283-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd6e7f7b6a224cac8b8e52bf19032911", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "edd47158-6f4b-44a1-8e82-0411205ad299", "external-id": "nsx-vlan-transportzone-587", "segmentation_id": 587, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap990aae71-9e", "ovs_interfaceid": "990aae71-9e8e-4da6-9e57-5c9257d127a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 832.893672] env[65522]: DEBUG oslo_vmware.api [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114066, 'name': ResetVM_Task, 'duration_secs': 0.117381} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.894035] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Did hard reboot of VM {{(pid=65522) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 832.894162] env[65522]: DEBUG nova.compute.manager [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 832.894990] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9407faba-3270-4e68-a68b-f80169724c00 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.031098] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f79808-ce3d-45a2-a65b-f7c32ba986f9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.037087] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Releasing lock "refresh_cache-18568baa-83de-4b54-addf-dee66211c8d7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 833.037557] env[65522]: DEBUG nova.compute.manager [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Instance network_info: |[{"id": "990aae71-9e8e-4da6-9e57-5c9257d127a4", "address": "fa:16:3e:1b:6a:ed", "network": {"id": "5165f03c-e790-4121-b5d7-3c5176afdd7e", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-551605283-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd6e7f7b6a224cac8b8e52bf19032911", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "edd47158-6f4b-44a1-8e82-0411205ad299", "external-id": "nsx-vlan-transportzone-587", "segmentation_id": 587, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap990aae71-9e", "ovs_interfaceid": "990aae71-9e8e-4da6-9e57-5c9257d127a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 833.040618] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:6a:ed', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'edd47158-6f4b-44a1-8e82-0411205ad299', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '990aae71-9e8e-4da6-9e57-5c9257d127a4', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.050476] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Creating folder: Project (fd6e7f7b6a224cac8b8e52bf19032911). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.050963] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-510a568f-82b4-4dcc-8894-c47de0281653 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.054133] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10928bf8-33b6-46a2-8552-ece2c0034744 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.087316] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c339e5-662a-4529-a140-f4a40d1503a7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.095898] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ab5010-a8ce-4c7d-8906-5be52a4dc49f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.101793] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Created folder: Project (fd6e7f7b6a224cac8b8e52bf19032911) in parent group-v994660. [ 833.101999] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Creating folder: Instances. Parent ref: group-v994803. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.102722] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-39b54a8a-eb18-489f-8748-169c4c1c81e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.112243] env[65522]: DEBUG nova.compute.provider_tree [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 833.125939] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Created folder: Instances in parent group-v994803. [ 833.126074] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 833.126274] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 833.126558] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b04e55f-49d5-4e2b-8df3-361a37e85680 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.147397] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.147397] env[65522]: value = "task-5114069" [ 833.147397] env[65522]: _type = "Task" [ 833.147397] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.157188] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114069, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.250450] env[65522]: DEBUG nova.compute.manager [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Received event network-changed-990aae71-9e8e-4da6-9e57-5c9257d127a4 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 833.250882] env[65522]: DEBUG nova.compute.manager [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Refreshing instance network info cache due to event network-changed-990aae71-9e8e-4da6-9e57-5c9257d127a4. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 833.251377] env[65522]: DEBUG oslo_concurrency.lockutils [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] Acquiring lock "refresh_cache-18568baa-83de-4b54-addf-dee66211c8d7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.251460] env[65522]: DEBUG oslo_concurrency.lockutils [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] Acquired lock "refresh_cache-18568baa-83de-4b54-addf-dee66211c8d7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 833.251708] env[65522]: DEBUG nova.network.neutron [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Refreshing network info cache for port 990aae71-9e8e-4da6-9e57-5c9257d127a4 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 833.411977] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8339824a-d4ff-4683-b717-44ac9a3b7a36 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "7c6f8218-602d-44f3-8012-de5a96972785" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.714s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 833.454966] env[65522]: DEBUG nova.compute.manager [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 833.480295] env[65522]: DEBUG nova.virt.hardware [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 833.480575] env[65522]: DEBUG nova.virt.hardware [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 833.480773] env[65522]: DEBUG nova.virt.hardware [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 833.481011] env[65522]: DEBUG nova.virt.hardware [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 833.481202] env[65522]: DEBUG nova.virt.hardware [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 833.481391] env[65522]: DEBUG nova.virt.hardware [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 833.481908] env[65522]: DEBUG nova.virt.hardware [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.481908] env[65522]: DEBUG nova.virt.hardware [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 833.482101] env[65522]: DEBUG nova.virt.hardware [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 833.482317] env[65522]: DEBUG nova.virt.hardware [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 833.482545] env[65522]: DEBUG nova.virt.hardware [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 833.483588] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a019e6ec-6611-4498-9cd3-433ad373cbfe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.493675] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ccd2fe-0ac0-4379-8ac5-76d664d7df92 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.511031] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 833.511142] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994802', 'volume_id': 'ec224e94-bcce-4815-96c0-d6f0d022ccbb', 'name': 'volume-ec224e94-bcce-4815-96c0-d6f0d022ccbb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20', 'attached_at': '', 'detached_at': '', 'volume_id': 'ec224e94-bcce-4815-96c0-d6f0d022ccbb', 'serial': 'ec224e94-bcce-4815-96c0-d6f0d022ccbb'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 833.512008] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c82eb3-76a3-4feb-ad2c-3676807ea28d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.529931] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee32f33-06ec-48c8-8c70-1dc9d2b015d0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.556734] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Reconfiguring VM instance instance-0000000c to attach disk [datastore1] volume-ec224e94-bcce-4815-96c0-d6f0d022ccbb/volume-ec224e94-bcce-4815-96c0-d6f0d022ccbb.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.557150] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51d28dd8-71f1-4299-b570-8f664eade9dc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.577180] env[65522]: DEBUG oslo_vmware.api [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Waiting for the task: (returnval){ [ 833.577180] env[65522]: value = "task-5114070" [ 833.577180] env[65522]: _type = "Task" [ 833.577180] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.586616] env[65522]: DEBUG oslo_vmware.api [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Task: {'id': task-5114070, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.636017] env[65522]: ERROR nova.scheduler.client.report [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [req-6b823498-c3fb-4f8c-ba40-182343140589] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9a0d1ac8-ed26-4657-8899-fbceab29db0d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6b823498-c3fb-4f8c-ba40-182343140589"}]} [ 833.658128] env[65522]: DEBUG oslo_vmware.rw_handles [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d401b6-91b6-02f5-05dc-d7ec890d3309/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 833.659928] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea4e7a9-3f87-4359-b2ff-4edb68426192 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.663283] env[65522]: DEBUG nova.scheduler.client.report [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Refreshing inventories for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 833.668041] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114069, 'name': CreateVM_Task, 'duration_secs': 0.452794} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.669162] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 833.669562] env[65522]: WARNING neutronclient.v2_0.client [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.669943] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.670101] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 833.670411] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 833.672037] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8874f29f-6fe5-4758-9a15-4f3015bf4ed8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.673785] env[65522]: DEBUG oslo_vmware.rw_handles [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d401b6-91b6-02f5-05dc-d7ec890d3309/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 833.674018] env[65522]: ERROR oslo_vmware.rw_handles [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d401b6-91b6-02f5-05dc-d7ec890d3309/disk-0.vmdk due to incomplete transfer. [ 833.674600] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-26d4a849-4e4f-450a-81e2-f4f0e9639573 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.677450] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Waiting for the task: (returnval){ [ 833.677450] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526ea704-1535-845a-da7a-173d2708d07e" [ 833.677450] env[65522]: _type = "Task" [ 833.677450] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.684212] env[65522]: DEBUG oslo_vmware.rw_handles [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d401b6-91b6-02f5-05dc-d7ec890d3309/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 833.684212] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Uploaded image eb493893-0d43-4227-a4d0-f0c68d713fef to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 833.686243] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 833.687465] env[65522]: DEBUG nova.scheduler.client.report [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Updating ProviderTree inventory for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 833.687465] env[65522]: DEBUG nova.compute.provider_tree [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 833.692606] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-cd259de0-ff1f-40c3-941b-a1761539fe3c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.694185] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526ea704-1535-845a-da7a-173d2708d07e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.700660] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 833.700660] env[65522]: value = "task-5114071" [ 833.700660] env[65522]: _type = "Task" [ 833.700660] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.711168] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114071, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.712389] env[65522]: DEBUG nova.scheduler.client.report [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Refreshing aggregate associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, aggregates: None {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 833.733805] env[65522]: DEBUG nova.scheduler.client.report [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Refreshing trait associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 833.755592] env[65522]: WARNING neutronclient.v2_0.client [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.756315] env[65522]: WARNING openstack [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 833.756670] env[65522]: WARNING openstack [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 833.775234] env[65522]: DEBUG nova.compute.manager [req-fe51011e-38da-43ea-806a-c33fd08ed57f req-46131032-223a-4791-9080-7362c5ded3c8 service nova] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Received event network-vif-plugged-6f040560-31cf-4beb-916d-44d19ed60c9e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 833.775363] env[65522]: DEBUG oslo_concurrency.lockutils [req-fe51011e-38da-43ea-806a-c33fd08ed57f req-46131032-223a-4791-9080-7362c5ded3c8 service nova] Acquiring lock "0b48b747-149c-4163-bdd0-7d0b07ddacfd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.775540] env[65522]: DEBUG oslo_concurrency.lockutils [req-fe51011e-38da-43ea-806a-c33fd08ed57f req-46131032-223a-4791-9080-7362c5ded3c8 service nova] Lock "0b48b747-149c-4163-bdd0-7d0b07ddacfd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.775810] env[65522]: DEBUG oslo_concurrency.lockutils [req-fe51011e-38da-43ea-806a-c33fd08ed57f req-46131032-223a-4791-9080-7362c5ded3c8 service nova] Lock "0b48b747-149c-4163-bdd0-7d0b07ddacfd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 833.775953] env[65522]: DEBUG nova.compute.manager [req-fe51011e-38da-43ea-806a-c33fd08ed57f req-46131032-223a-4791-9080-7362c5ded3c8 service nova] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] No waiting events found dispatching network-vif-plugged-6f040560-31cf-4beb-916d-44d19ed60c9e {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 833.776921] env[65522]: WARNING nova.compute.manager [req-fe51011e-38da-43ea-806a-c33fd08ed57f req-46131032-223a-4791-9080-7362c5ded3c8 service nova] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Received unexpected event network-vif-plugged-6f040560-31cf-4beb-916d-44d19ed60c9e for instance with vm_state building and task_state spawning. [ 833.907465] env[65522]: DEBUG nova.network.neutron [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Successfully updated port: 6f040560-31cf-4beb-916d-44d19ed60c9e {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 833.958552] env[65522]: WARNING openstack [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 833.959622] env[65522]: WARNING openstack [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 834.051255] env[65522]: WARNING neutronclient.v2_0.client [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 834.052114] env[65522]: WARNING openstack [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 834.053259] env[65522]: WARNING openstack [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 834.088049] env[65522]: DEBUG oslo_vmware.api [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Task: {'id': task-5114070, 'name': ReconfigVM_Task, 'duration_secs': 0.462556} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.088370] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Reconfigured VM instance instance-0000000c to attach disk [datastore1] volume-ec224e94-bcce-4815-96c0-d6f0d022ccbb/volume-ec224e94-bcce-4815-96c0-d6f0d022ccbb.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 834.095866] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-884aebd4-3613-46a6-9120-e92182cd139f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.116406] env[65522]: DEBUG oslo_vmware.api [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Waiting for the task: (returnval){ [ 834.116406] env[65522]: value = "task-5114072" [ 834.116406] env[65522]: _type = "Task" [ 834.116406] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.129209] env[65522]: DEBUG oslo_vmware.api [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Task: {'id': task-5114072, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.154343] env[65522]: DEBUG nova.network.neutron [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Updated VIF entry in instance network info cache for port 990aae71-9e8e-4da6-9e57-5c9257d127a4. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 834.154726] env[65522]: DEBUG nova.network.neutron [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Updating instance_info_cache with network_info: [{"id": "990aae71-9e8e-4da6-9e57-5c9257d127a4", "address": "fa:16:3e:1b:6a:ed", "network": {"id": "5165f03c-e790-4121-b5d7-3c5176afdd7e", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-551605283-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd6e7f7b6a224cac8b8e52bf19032911", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "edd47158-6f4b-44a1-8e82-0411205ad299", "external-id": "nsx-vlan-transportzone-587", "segmentation_id": 587, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap990aae71-9e", "ovs_interfaceid": "990aae71-9e8e-4da6-9e57-5c9257d127a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 834.190829] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526ea704-1535-845a-da7a-173d2708d07e, 'name': SearchDatastore_Task, 'duration_secs': 0.011791} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.191151] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 834.191440] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.191698] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.191847] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 834.193028] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.193028] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f663f6b6-6b27-4d05-a991-edeaeb93743b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.205991] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.206284] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 834.207615] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9169dad8-edf7-4d7d-8e4c-734f4ccbd679 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.212694] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114071, 'name': Destroy_Task, 'duration_secs': 0.383046} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.215712] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Destroyed the VM [ 834.215963] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 834.216424] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9c5fb2cf-64cb-49fd-97d3-cfda377f941f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.219415] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Waiting for the task: (returnval){ [ 834.219415] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e1b69a-101c-b983-64d1-1243b87e8827" [ 834.219415] env[65522]: _type = "Task" [ 834.219415] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.225968] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 834.225968] env[65522]: value = "task-5114073" [ 834.225968] env[65522]: _type = "Task" [ 834.225968] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.229051] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e1b69a-101c-b983-64d1-1243b87e8827, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.245119] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114073, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.390141] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-164d1b27-8de7-4dc1-97ce-e5df91f6eb55 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.398052] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb30541e-4a5a-4e73-b0a8-45771ad80e7e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.431036] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "refresh_cache-0b48b747-149c-4163-bdd0-7d0b07ddacfd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.431208] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "refresh_cache-0b48b747-149c-4163-bdd0-7d0b07ddacfd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 834.431397] env[65522]: DEBUG nova.network.neutron [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 834.434607] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29012795-109f-43a0-b9d2-78e3cf313ede {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.444084] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e458a4a0-4ca5-42cd-a266-f83e8bfee2d0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.460060] env[65522]: DEBUG nova.compute.provider_tree [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.629767] env[65522]: DEBUG oslo_vmware.api [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Task: {'id': task-5114072, 'name': ReconfigVM_Task, 'duration_secs': 0.168492} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.630012] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994802', 'volume_id': 'ec224e94-bcce-4815-96c0-d6f0d022ccbb', 'name': 'volume-ec224e94-bcce-4815-96c0-d6f0d022ccbb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20', 'attached_at': '', 'detached_at': '', 'volume_id': 'ec224e94-bcce-4815-96c0-d6f0d022ccbb', 'serial': 'ec224e94-bcce-4815-96c0-d6f0d022ccbb'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 834.659970] env[65522]: DEBUG oslo_concurrency.lockutils [req-e10b65f3-e526-4277-ade6-038e5f14cad5 req-73a56c1f-c5c6-43bb-a87c-3d9b419563a7 service nova] Releasing lock "refresh_cache-18568baa-83de-4b54-addf-dee66211c8d7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 834.731916] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e1b69a-101c-b983-64d1-1243b87e8827, 'name': SearchDatastore_Task, 'duration_secs': 0.014051} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.735934] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08d6251d-1618-4919-a389-b1d185224bb4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.742696] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Waiting for the task: (returnval){ [ 834.742696] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5245cd7d-8365-1af2-dfdb-112a662bbc3a" [ 834.742696] env[65522]: _type = "Task" [ 834.742696] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.746269] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114073, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.755395] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5245cd7d-8365-1af2-dfdb-112a662bbc3a, 'name': SearchDatastore_Task, 'duration_secs': 0.010416} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.755670] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 834.755931] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 18568baa-83de-4b54-addf-dee66211c8d7/18568baa-83de-4b54-addf-dee66211c8d7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 834.756239] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ab1afea8-d8b5-46ec-971d-82a12e342e30 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.764111] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Waiting for the task: (returnval){ [ 834.764111] env[65522]: value = "task-5114074" [ 834.764111] env[65522]: _type = "Task" [ 834.764111] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.773395] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114074, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.935901] env[65522]: WARNING openstack [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 834.936479] env[65522]: WARNING openstack [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 834.963931] env[65522]: DEBUG nova.scheduler.client.report [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 834.977671] env[65522]: DEBUG nova.network.neutron [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 835.000857] env[65522]: WARNING openstack [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 835.001455] env[65522]: WARNING openstack [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 835.072267] env[65522]: WARNING neutronclient.v2_0.client [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 835.073078] env[65522]: WARNING openstack [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 835.073553] env[65522]: WARNING openstack [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 835.166304] env[65522]: DEBUG nova.network.neutron [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Updating instance_info_cache with network_info: [{"id": "6f040560-31cf-4beb-916d-44d19ed60c9e", "address": "fa:16:3e:ee:fb:1b", "network": {"id": "dc9d8462-3539-4cc5-b21a-e1136a8df8e6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1580876213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b5c14a856d454048856539895f220c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f040560-31", "ovs_interfaceid": "6f040560-31cf-4beb-916d-44d19ed60c9e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 835.245923] env[65522]: DEBUG oslo_vmware.api [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114073, 'name': RemoveSnapshot_Task, 'duration_secs': 0.53599} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.246346] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 835.246682] env[65522]: INFO nova.compute.manager [None req-139c6edb-af56-45cf-a6cd-3ea71b7f4867 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Took 14.40 seconds to snapshot the instance on the hypervisor. [ 835.275103] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114074, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.469848] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.047s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.470499] env[65522]: DEBUG nova.compute.manager [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 835.473912] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.718s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.475653] env[65522]: INFO nova.compute.claims [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 835.669880] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "refresh_cache-0b48b747-149c-4163-bdd0-7d0b07ddacfd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 835.670396] env[65522]: DEBUG nova.compute.manager [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Instance network_info: |[{"id": "6f040560-31cf-4beb-916d-44d19ed60c9e", "address": "fa:16:3e:ee:fb:1b", "network": {"id": "dc9d8462-3539-4cc5-b21a-e1136a8df8e6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1580876213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b5c14a856d454048856539895f220c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f040560-31", "ovs_interfaceid": "6f040560-31cf-4beb-916d-44d19ed60c9e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 835.670990] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:fb:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6f040560-31cf-4beb-916d-44d19ed60c9e', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.682943] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 835.684979] env[65522]: DEBUG nova.objects.instance [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Lazy-loading 'flavor' on Instance uuid 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.685714] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 835.686767] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4379b64-61ee-4d18-911c-21f285d24d24 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.716498] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.716498] env[65522]: value = "task-5114075" [ 835.716498] env[65522]: _type = "Task" [ 835.716498] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.728796] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114075, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.781236] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114074, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516534} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.781821] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 18568baa-83de-4b54-addf-dee66211c8d7/18568baa-83de-4b54-addf-dee66211c8d7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 835.782135] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 835.782539] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-832c6b0d-5301-45ff-a172-b205962f9509 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.792592] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Waiting for the task: (returnval){ [ 835.792592] env[65522]: value = "task-5114076" [ 835.792592] env[65522]: _type = "Task" [ 835.792592] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.810170] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114076, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.976535] env[65522]: DEBUG nova.compute.utils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 835.979768] env[65522]: DEBUG nova.compute.manager [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 835.979927] env[65522]: DEBUG nova.network.neutron [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 835.980268] env[65522]: WARNING neutronclient.v2_0.client [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 835.980599] env[65522]: WARNING neutronclient.v2_0.client [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 835.981209] env[65522]: WARNING openstack [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 835.981615] env[65522]: WARNING openstack [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 835.995677] env[65522]: DEBUG nova.compute.manager [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Received event network-changed-6f040560-31cf-4beb-916d-44d19ed60c9e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 835.995979] env[65522]: DEBUG nova.compute.manager [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Refreshing instance network info cache due to event network-changed-6f040560-31cf-4beb-916d-44d19ed60c9e. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 835.996214] env[65522]: DEBUG oslo_concurrency.lockutils [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] Acquiring lock "refresh_cache-0b48b747-149c-4163-bdd0-7d0b07ddacfd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.996370] env[65522]: DEBUG oslo_concurrency.lockutils [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] Acquired lock "refresh_cache-0b48b747-149c-4163-bdd0-7d0b07ddacfd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 835.996500] env[65522]: DEBUG nova.network.neutron [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Refreshing network info cache for port 6f040560-31cf-4beb-916d-44d19ed60c9e {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 836.048347] env[65522]: DEBUG nova.policy [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '421f99d51af04aefbc077c8d6756a8aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9236eeef7e5b4f5fbee26ad7fede9106', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 836.191456] env[65522]: DEBUG oslo_concurrency.lockutils [None req-607d6e50-7bf6-452e-ac55-caa5d81ecfa5 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.311s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 836.230146] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114075, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.304929] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114076, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079959} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.305227] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.306081] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c06c38-a6ea-4ed6-8849-e6093465667d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.344567] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] 18568baa-83de-4b54-addf-dee66211c8d7/18568baa-83de-4b54-addf-dee66211c8d7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.344966] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4ca78a4-6a65-4def-9a4f-2aeae6571136 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.373633] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Waiting for the task: (returnval){ [ 836.373633] env[65522]: value = "task-5114077" [ 836.373633] env[65522]: _type = "Task" [ 836.373633] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.384793] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114077, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.480815] env[65522]: DEBUG nova.compute.manager [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 836.503746] env[65522]: WARNING neutronclient.v2_0.client [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 836.503746] env[65522]: WARNING openstack [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.503746] env[65522]: WARNING openstack [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.668895] env[65522]: WARNING openstack [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.671707] env[65522]: WARNING openstack [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.729809] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114075, 'name': CreateVM_Task, 'duration_secs': 0.782205} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.732711] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 836.733468] env[65522]: WARNING neutronclient.v2_0.client [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 836.733848] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.734864] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 836.734864] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 836.734864] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4bf733d-f2f4-463e-803e-b11d42c1fa0e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.740877] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 836.740877] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52808921-104d-dcee-c184-0728f0280e72" [ 836.740877] env[65522]: _type = "Task" [ 836.740877] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.742224] env[65522]: WARNING neutronclient.v2_0.client [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 836.742296] env[65522]: WARNING openstack [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.742640] env[65522]: WARNING openstack [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.763334] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52808921-104d-dcee-c184-0728f0280e72, 'name': SearchDatastore_Task, 'duration_secs': 0.011429} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.763663] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 836.763910] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.764188] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.764334] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 836.764903] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.764903] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-edce48b2-4682-40c4-9fa0-db3f5866f5cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.778585] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.778585] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 836.778585] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e35e846-9b4c-40f6-b8ae-9b0db3e4ce56 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.785090] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 836.785090] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521220b0-5627-a385-4bf1-77198aabc445" [ 836.785090] env[65522]: _type = "Task" [ 836.785090] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.799910] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521220b0-5627-a385-4bf1-77198aabc445, 'name': SearchDatastore_Task, 'duration_secs': 0.012955} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.800953] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0afd6235-e9e4-4b66-b873-7fd3c81b78b6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.809032] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 836.809032] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c0985a-c697-a7d1-da2b-b0f26d160f72" [ 836.809032] env[65522]: _type = "Task" [ 836.809032] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.824929] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c0985a-c697-a7d1-da2b-b0f26d160f72, 'name': SearchDatastore_Task, 'duration_secs': 0.012084} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.824929] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 836.824929] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 0b48b747-149c-4163-bdd0-7d0b07ddacfd/0b48b747-149c-4163-bdd0-7d0b07ddacfd.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 836.825558] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-83c5ec6a-ec89-4f79-83cc-4cd6e9062723 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.844625] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 836.844625] env[65522]: value = "task-5114078" [ 836.844625] env[65522]: _type = "Task" [ 836.844625] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.851563] env[65522]: DEBUG nova.network.neutron [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Successfully created port: 20608ea3-7885-4006-82d3-66674cbd0e61 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 836.865723] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114078, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.884682] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114077, 'name': ReconfigVM_Task, 'duration_secs': 0.305076} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.887734] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Reconfigured VM instance instance-00000033 to attach disk [datastore1] 18568baa-83de-4b54-addf-dee66211c8d7/18568baa-83de-4b54-addf-dee66211c8d7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.890425] env[65522]: DEBUG nova.network.neutron [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Updated VIF entry in instance network info cache for port 6f040560-31cf-4beb-916d-44d19ed60c9e. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 836.890785] env[65522]: DEBUG nova.network.neutron [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Updating instance_info_cache with network_info: [{"id": "6f040560-31cf-4beb-916d-44d19ed60c9e", "address": "fa:16:3e:ee:fb:1b", "network": {"id": "dc9d8462-3539-4cc5-b21a-e1136a8df8e6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1580876213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b5c14a856d454048856539895f220c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6f040560-31", "ovs_interfaceid": "6f040560-31cf-4beb-916d-44d19ed60c9e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 836.892197] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e6c8dbad-71f9-4863-b7d8-78520dc275c2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.901723] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Waiting for the task: (returnval){ [ 836.901723] env[65522]: value = "task-5114079" [ 836.901723] env[65522]: _type = "Task" [ 836.901723] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.911929] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114079, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.946991] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Acquiring lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 836.947278] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 837.232770] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e333e7-bcd1-4b98-ae74-5cb2289db26f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.244889] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece489c5-8731-4a60-8973-1360bab8104d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.283089] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1e8d64-7819-49f9-8c29-e2ba5e243dff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.296073] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5c02f4-1f4a-4692-8c28-f0c985fdfd1e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.313216] env[65522]: DEBUG nova.compute.provider_tree [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.356055] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114078, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.395968] env[65522]: DEBUG oslo_concurrency.lockutils [req-fcac5898-1adc-4700-b819-925c8ab3dc26 req-a9f4cbfc-4e1b-4645-91cf-a766215bfce5 service nova] Releasing lock "refresh_cache-0b48b747-149c-4163-bdd0-7d0b07ddacfd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 837.412714] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114079, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.451167] env[65522]: INFO nova.compute.manager [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Detaching volume ec224e94-bcce-4815-96c0-d6f0d022ccbb [ 837.488908] env[65522]: INFO nova.virt.block_device [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Attempting to driver detach volume ec224e94-bcce-4815-96c0-d6f0d022ccbb from mountpoint /dev/sdb [ 837.489180] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 837.489375] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994802', 'volume_id': 'ec224e94-bcce-4815-96c0-d6f0d022ccbb', 'name': 'volume-ec224e94-bcce-4815-96c0-d6f0d022ccbb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20', 'attached_at': '', 'detached_at': '', 'volume_id': 'ec224e94-bcce-4815-96c0-d6f0d022ccbb', 'serial': 'ec224e94-bcce-4815-96c0-d6f0d022ccbb'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 837.490370] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a490d98-3350-41b4-874a-84982e65cdfa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.494729] env[65522]: DEBUG nova.compute.manager [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 837.518421] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a81c47f-5f80-403c-8dea-268be5337538 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.528860] env[65522]: DEBUG nova.virt.hardware [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 837.529119] env[65522]: DEBUG nova.virt.hardware [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 837.529278] env[65522]: DEBUG nova.virt.hardware [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 837.529516] env[65522]: DEBUG nova.virt.hardware [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 837.529644] env[65522]: DEBUG nova.virt.hardware [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 837.529790] env[65522]: DEBUG nova.virt.hardware [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 837.530007] env[65522]: DEBUG nova.virt.hardware [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 837.530176] env[65522]: DEBUG nova.virt.hardware [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 837.530345] env[65522]: DEBUG nova.virt.hardware [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 837.530509] env[65522]: DEBUG nova.virt.hardware [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 837.530691] env[65522]: DEBUG nova.virt.hardware [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 837.531515] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16190786-da7d-41f0-a17e-30d210120e9a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.534650] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7870a531-eea0-4831-b955-e9f7612cc298 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.563384] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03c877c-832d-4261-86fe-65feec6fe607 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.568286] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074e00ac-e500-49e4-aa36-99b2decd40a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.585449] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] The volume has not been displaced from its original location: [datastore1] volume-ec224e94-bcce-4815-96c0-d6f0d022ccbb/volume-ec224e94-bcce-4815-96c0-d6f0d022ccbb.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 837.591205] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Reconfiguring VM instance instance-0000000c to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 837.601041] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83dd8f90-1ee6-4f80-ac5c-21a74b979825 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.620628] env[65522]: DEBUG oslo_vmware.api [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Waiting for the task: (returnval){ [ 837.620628] env[65522]: value = "task-5114080" [ 837.620628] env[65522]: _type = "Task" [ 837.620628] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.629626] env[65522]: DEBUG oslo_vmware.api [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Task: {'id': task-5114080, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.817458] env[65522]: DEBUG nova.scheduler.client.report [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 837.856196] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114078, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518864} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.856517] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 0b48b747-149c-4163-bdd0-7d0b07ddacfd/0b48b747-149c-4163-bdd0-7d0b07ddacfd.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 837.856733] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.857129] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a90e443c-888c-4a95-8af2-a86553b61cb0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.865653] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 837.865653] env[65522]: value = "task-5114081" [ 837.865653] env[65522]: _type = "Task" [ 837.865653] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.876944] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114081, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.913509] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114079, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.132097] env[65522]: DEBUG oslo_vmware.api [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Task: {'id': task-5114080, 'name': ReconfigVM_Task, 'duration_secs': 0.256798} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.132379] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Reconfigured VM instance instance-0000000c to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 838.137424] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a79e2369-aeae-42d7-996c-56ba3cf7525d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.154560] env[65522]: DEBUG oslo_vmware.api [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Waiting for the task: (returnval){ [ 838.154560] env[65522]: value = "task-5114082" [ 838.154560] env[65522]: _type = "Task" [ 838.154560] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.165701] env[65522]: DEBUG oslo_vmware.api [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Task: {'id': task-5114082, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.326437] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.852s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.326965] env[65522]: DEBUG nova.compute.manager [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 838.330013] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.566s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.330269] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.332703] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 36.443s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.373768] env[65522]: INFO nova.scheduler.client.report [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Deleted allocations for instance b2d70982-54e7-459c-a0d8-48bf4b6e4345 [ 838.383550] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114081, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074674} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.383815] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.387199] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684458a9-1e5f-44ae-950d-2ff559d91315 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.413620] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 0b48b747-149c-4163-bdd0-7d0b07ddacfd/0b48b747-149c-4163-bdd0-7d0b07ddacfd.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.417172] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd7066e6-ab65-44b2-81b9-7bfa4b003ca2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.439688] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114079, 'name': Rename_Task, 'duration_secs': 1.169337} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.440815] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 838.441216] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 838.441216] env[65522]: value = "task-5114083" [ 838.441216] env[65522]: _type = "Task" [ 838.441216] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.441428] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-127ecb4b-afae-48d8-b415-fa039b583489 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.454547] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114083, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.456246] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Waiting for the task: (returnval){ [ 838.456246] env[65522]: value = "task-5114084" [ 838.456246] env[65522]: _type = "Task" [ 838.456246] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.465966] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114084, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.552530] env[65522]: DEBUG nova.compute.manager [req-fa6fa1ac-1e48-42e3-b30b-7342f0b15909 req-cd5e1d52-7651-481d-b9d5-7a686a0dab5f service nova] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Received event network-vif-plugged-20608ea3-7885-4006-82d3-66674cbd0e61 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 838.552748] env[65522]: DEBUG oslo_concurrency.lockutils [req-fa6fa1ac-1e48-42e3-b30b-7342f0b15909 req-cd5e1d52-7651-481d-b9d5-7a686a0dab5f service nova] Acquiring lock "880390f5-00f1-4dea-9345-29c06289b340-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.552911] env[65522]: DEBUG oslo_concurrency.lockutils [req-fa6fa1ac-1e48-42e3-b30b-7342f0b15909 req-cd5e1d52-7651-481d-b9d5-7a686a0dab5f service nova] Lock "880390f5-00f1-4dea-9345-29c06289b340-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.555793] env[65522]: DEBUG oslo_concurrency.lockutils [req-fa6fa1ac-1e48-42e3-b30b-7342f0b15909 req-cd5e1d52-7651-481d-b9d5-7a686a0dab5f service nova] Lock "880390f5-00f1-4dea-9345-29c06289b340-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.555986] env[65522]: DEBUG nova.compute.manager [req-fa6fa1ac-1e48-42e3-b30b-7342f0b15909 req-cd5e1d52-7651-481d-b9d5-7a686a0dab5f service nova] [instance: 880390f5-00f1-4dea-9345-29c06289b340] No waiting events found dispatching network-vif-plugged-20608ea3-7885-4006-82d3-66674cbd0e61 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 838.556187] env[65522]: WARNING nova.compute.manager [req-fa6fa1ac-1e48-42e3-b30b-7342f0b15909 req-cd5e1d52-7651-481d-b9d5-7a686a0dab5f service nova] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Received unexpected event network-vif-plugged-20608ea3-7885-4006-82d3-66674cbd0e61 for instance with vm_state building and task_state spawning. [ 838.559221] env[65522]: DEBUG nova.network.neutron [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Successfully updated port: 20608ea3-7885-4006-82d3-66674cbd0e61 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 838.562051] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "da211708-a6a2-4e03-bb41-c5b93564de20" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.562405] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "da211708-a6a2-4e03-bb41-c5b93564de20" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.562511] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "da211708-a6a2-4e03-bb41-c5b93564de20-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.562920] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "da211708-a6a2-4e03-bb41-c5b93564de20-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.563185] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "da211708-a6a2-4e03-bb41-c5b93564de20-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.565107] env[65522]: INFO nova.compute.manager [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Terminating instance [ 838.666099] env[65522]: DEBUG oslo_vmware.api [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Task: {'id': task-5114082, 'name': ReconfigVM_Task, 'duration_secs': 0.157202} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.666099] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994802', 'volume_id': 'ec224e94-bcce-4815-96c0-d6f0d022ccbb', 'name': 'volume-ec224e94-bcce-4815-96c0-d6f0d022ccbb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20', 'attached_at': '', 'detached_at': '', 'volume_id': 'ec224e94-bcce-4815-96c0-d6f0d022ccbb', 'serial': 'ec224e94-bcce-4815-96c0-d6f0d022ccbb'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 838.837055] env[65522]: DEBUG nova.compute.utils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 838.841371] env[65522]: INFO nova.compute.claims [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.846725] env[65522]: DEBUG nova.compute.manager [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 838.847245] env[65522]: DEBUG nova.network.neutron [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 838.847490] env[65522]: WARNING neutronclient.v2_0.client [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.847900] env[65522]: WARNING neutronclient.v2_0.client [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.849164] env[65522]: WARNING openstack [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.849615] env[65522]: WARNING openstack [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.881964] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2052ecde-63d7-4f6e-92bf-b66cdb73051f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "b2d70982-54e7-459c-a0d8-48bf4b6e4345" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.440s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.957866] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114083, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.966578] env[65522]: DEBUG nova.policy [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7ef25c2fe4643748839c6fc67e74ad0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37fc01876567476f9b93d765b2cfddc8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 838.973340] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114084, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.065761] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Acquiring lock "refresh_cache-880390f5-00f1-4dea-9345-29c06289b340" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.066178] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Acquired lock "refresh_cache-880390f5-00f1-4dea-9345-29c06289b340" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 839.066178] env[65522]: DEBUG nova.network.neutron [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 839.071519] env[65522]: DEBUG nova.compute.manager [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 839.071809] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 839.072783] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b8991d-c663-4f16-af4f-e37b7113c6dd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.082852] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 839.084701] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d6a479a-4fc4-466a-8837-971a47067172 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.094206] env[65522]: DEBUG oslo_vmware.api [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 839.094206] env[65522]: value = "task-5114085" [ 839.094206] env[65522]: _type = "Task" [ 839.094206] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.105616] env[65522]: DEBUG oslo_vmware.api [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114085, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.238184] env[65522]: DEBUG nova.objects.instance [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Lazy-loading 'flavor' on Instance uuid 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 839.348724] env[65522]: INFO nova.compute.resource_tracker [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating resource usage from migration 98e33664-f395-480b-a10d-3904a0963458 [ 839.359412] env[65522]: DEBUG nova.network.neutron [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Successfully created port: cb206964-4331-495a-94ce-c2b95dccc600 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 839.362074] env[65522]: DEBUG nova.compute.manager [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 839.457458] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114083, 'name': ReconfigVM_Task, 'duration_secs': 0.73078} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.458193] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 0b48b747-149c-4163-bdd0-7d0b07ddacfd/0b48b747-149c-4163-bdd0-7d0b07ddacfd.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.461813] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8fcf4aa3-53f4-4728-ba19-8c077455fb31 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.472333] env[65522]: DEBUG oslo_vmware.api [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114084, 'name': PowerOnVM_Task, 'duration_secs': 0.52163} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.473771] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 839.473982] env[65522]: INFO nova.compute.manager [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Took 8.69 seconds to spawn the instance on the hypervisor. [ 839.474222] env[65522]: DEBUG nova.compute.manager [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 839.474592] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 839.474592] env[65522]: value = "task-5114086" [ 839.474592] env[65522]: _type = "Task" [ 839.474592] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.475345] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40435893-69b2-412e-b183-33761b65bff4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.495735] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114086, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.568954] env[65522]: WARNING openstack [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 839.569321] env[65522]: WARNING openstack [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 839.604759] env[65522]: DEBUG oslo_vmware.api [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114085, 'name': PowerOffVM_Task, 'duration_secs': 0.259233} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.604937] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 839.605113] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 839.605389] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8636f325-5226-488b-9e77-2962d10ed412 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.667588] env[65522]: DEBUG nova.network.neutron [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 839.703370] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 839.703588] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 839.703762] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Deleting the datastore file [datastore1] da211708-a6a2-4e03-bb41-c5b93564de20 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.704057] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-307b7de4-629f-45e0-8427-a1df6aa66122 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.715197] env[65522]: DEBUG oslo_vmware.api [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 839.715197] env[65522]: value = "task-5114088" [ 839.715197] env[65522]: _type = "Task" [ 839.715197] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.725297] env[65522]: DEBUG oslo_vmware.api [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114088, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.994839] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114086, 'name': Rename_Task, 'duration_secs': 0.153251} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.997793] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 840.002626] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7b24d4e2-d9dd-48b4-aa22-22348617a712 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.008757] env[65522]: INFO nova.compute.manager [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Took 51.46 seconds to build instance. [ 840.016646] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 840.016646] env[65522]: value = "task-5114089" [ 840.016646] env[65522]: _type = "Task" [ 840.016646] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.027798] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad3324f-c4fe-4e7e-a2d1-b2bf7ad5557a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.035724] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114089, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.046118] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a102df7-2d77-4015-b973-f16c13d5a3a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.086311] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d391b69f-fc83-4580-8327-376c61baa800 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.098407] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-439a3e58-7cda-4139-a237-5100d9c4d111 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.114285] env[65522]: DEBUG nova.compute.provider_tree [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.230376] env[65522]: DEBUG oslo_vmware.api [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114088, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.278957} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.230741] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.231010] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 840.231382] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 840.231632] env[65522]: INFO nova.compute.manager [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Took 1.16 seconds to destroy the instance on the hypervisor. [ 840.232037] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 840.232331] env[65522]: DEBUG nova.compute.manager [-] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 840.232470] env[65522]: DEBUG nova.network.neutron [-] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 840.232808] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 840.233586] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 840.233987] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 840.247664] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27c2ffb6-66c6-491b-9442-1c90eb49ad27 tempest-VolumesAssistedSnapshotsTest-2125981362 tempest-VolumesAssistedSnapshotsTest-2125981362-project-admin] Lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.300s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 840.293553] env[65522]: WARNING openstack [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 840.293956] env[65522]: WARNING openstack [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 840.369446] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 840.376734] env[65522]: DEBUG nova.compute.manager [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 840.407250] env[65522]: DEBUG nova.virt.hardware [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 840.407250] env[65522]: DEBUG nova.virt.hardware [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 840.407250] env[65522]: DEBUG nova.virt.hardware [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 840.407705] env[65522]: DEBUG nova.virt.hardware [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 840.408588] env[65522]: DEBUG nova.virt.hardware [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 840.408588] env[65522]: DEBUG nova.virt.hardware [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 840.408588] env[65522]: DEBUG nova.virt.hardware [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 840.410908] env[65522]: DEBUG nova.virt.hardware [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 840.410908] env[65522]: DEBUG nova.virt.hardware [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 840.410908] env[65522]: DEBUG nova.virt.hardware [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 840.410908] env[65522]: DEBUG nova.virt.hardware [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 840.410908] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eddcb43-8b47-445e-b788-1fcacf6d73e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.421754] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-294e9617-d0d6-4ba3-b749-143db95aa956 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.514431] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8df2ab5f-6c6a-4fa3-9d42-711232156071 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Lock "18568baa-83de-4b54-addf-dee66211c8d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.981s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 840.529502] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114089, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.618105] env[65522]: DEBUG nova.scheduler.client.report [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 840.655989] env[65522]: WARNING neutronclient.v2_0.client [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 840.656921] env[65522]: WARNING openstack [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 840.657188] env[65522]: WARNING openstack [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 840.765359] env[65522]: DEBUG nova.network.neutron [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Updating instance_info_cache with network_info: [{"id": "20608ea3-7885-4006-82d3-66674cbd0e61", "address": "fa:16:3e:94:61:f8", "network": {"id": "8033426d-3449-4534-bb51-86f8e1cbf396", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-324841709-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9236eeef7e5b4f5fbee26ad7fede9106", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20608ea3-78", "ovs_interfaceid": "20608ea3-7885-4006-82d3-66674cbd0e61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 841.029411] env[65522]: DEBUG oslo_vmware.api [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114089, 'name': PowerOnVM_Task, 'duration_secs': 0.847391} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.029819] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.029992] env[65522]: INFO nova.compute.manager [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Took 7.57 seconds to spawn the instance on the hypervisor. [ 841.030238] env[65522]: DEBUG nova.compute.manager [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 841.031087] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c643e30-0a62-459e-9518-2a01028fc335 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.101089] env[65522]: DEBUG nova.network.neutron [-] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 841.123359] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.790s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 841.123556] env[65522]: INFO nova.compute.manager [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Migrating [ 841.131417] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.240s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 841.135173] env[65522]: INFO nova.compute.claims [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 841.270451] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Releasing lock "refresh_cache-880390f5-00f1-4dea-9345-29c06289b340" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 841.270936] env[65522]: DEBUG nova.compute.manager [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Instance network_info: |[{"id": "20608ea3-7885-4006-82d3-66674cbd0e61", "address": "fa:16:3e:94:61:f8", "network": {"id": "8033426d-3449-4534-bb51-86f8e1cbf396", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-324841709-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9236eeef7e5b4f5fbee26ad7fede9106", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20608ea3-78", "ovs_interfaceid": "20608ea3-7885-4006-82d3-66674cbd0e61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 841.271608] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:61:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd829efb7-e98e-4b67-bd03-b0888287dbfd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '20608ea3-7885-4006-82d3-66674cbd0e61', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.281250] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Creating folder: Project (9236eeef7e5b4f5fbee26ad7fede9106). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 841.281713] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-18024c86-fca3-406b-9aa5-8981a7aac8d8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.296193] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Created folder: Project (9236eeef7e5b4f5fbee26ad7fede9106) in parent group-v994660. [ 841.296759] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Creating folder: Instances. Parent ref: group-v994807. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 841.297838] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bff7457d-3417-4c37-a34c-3e20fcd749a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.313760] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Created folder: Instances in parent group-v994807. [ 841.314273] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 841.314351] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 841.314629] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89ae9c69-ed14-4f00-97ee-ce5c573d4488 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.337057] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.337057] env[65522]: value = "task-5114093" [ 841.337057] env[65522]: _type = "Task" [ 841.337057] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.347543] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114093, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.394560] env[65522]: DEBUG nova.compute.manager [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Received event network-changed-20608ea3-7885-4006-82d3-66674cbd0e61 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 841.394829] env[65522]: DEBUG nova.compute.manager [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Refreshing instance network info cache due to event network-changed-20608ea3-7885-4006-82d3-66674cbd0e61. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 841.394965] env[65522]: DEBUG oslo_concurrency.lockutils [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] Acquiring lock "refresh_cache-880390f5-00f1-4dea-9345-29c06289b340" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.395238] env[65522]: DEBUG oslo_concurrency.lockutils [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] Acquired lock "refresh_cache-880390f5-00f1-4dea-9345-29c06289b340" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 841.395408] env[65522]: DEBUG nova.network.neutron [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Refreshing network info cache for port 20608ea3-7885-4006-82d3-66674cbd0e61 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 841.529225] env[65522]: DEBUG nova.compute.manager [req-314f009d-da6b-4028-8c36-a3fa67663fd2 req-ecd8f1be-4f1c-45e8-b1a6-bc0b430b8932 service nova] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Received event network-vif-deleted-345f682f-4aaa-4bc1-bb61-e9d25e6288c5 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 841.533405] env[65522]: DEBUG nova.network.neutron [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Successfully updated port: cb206964-4331-495a-94ce-c2b95dccc600 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 841.553322] env[65522]: INFO nova.compute.manager [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Took 47.46 seconds to build instance. [ 841.603194] env[65522]: INFO nova.compute.manager [-] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Took 1.37 seconds to deallocate network for instance. [ 841.653312] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.653312] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 841.653548] env[65522]: DEBUG nova.network.neutron [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 841.849611] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114093, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.898515] env[65522]: WARNING neutronclient.v2_0.client [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 841.899342] env[65522]: WARNING openstack [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 841.899858] env[65522]: WARNING openstack [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 842.037200] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.037426] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquired lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 842.037743] env[65522]: DEBUG nova.network.neutron [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 842.057770] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7138f3cc-d8a9-4318-ae66-bf7dd0423a2a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "0b48b747-149c-4163-bdd0-7d0b07ddacfd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.005s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 842.063339] env[65522]: WARNING openstack [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 842.063339] env[65522]: WARNING openstack [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 842.111094] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 842.150998] env[65522]: WARNING neutronclient.v2_0.client [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 842.151729] env[65522]: WARNING openstack [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 842.152141] env[65522]: WARNING openstack [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 842.160808] env[65522]: WARNING neutronclient.v2_0.client [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 842.161039] env[65522]: WARNING openstack [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 842.161380] env[65522]: WARNING openstack [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 842.369143] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 842.369143] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.370220] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114093, 'name': CreateVM_Task, 'duration_secs': 0.543402} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.374851] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 842.377489] env[65522]: DEBUG nova.network.neutron [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Updated VIF entry in instance network info cache for port 20608ea3-7885-4006-82d3-66674cbd0e61. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 842.377793] env[65522]: DEBUG nova.network.neutron [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Updating instance_info_cache with network_info: [{"id": "20608ea3-7885-4006-82d3-66674cbd0e61", "address": "fa:16:3e:94:61:f8", "network": {"id": "8033426d-3449-4534-bb51-86f8e1cbf396", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-324841709-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9236eeef7e5b4f5fbee26ad7fede9106", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20608ea3-78", "ovs_interfaceid": "20608ea3-7885-4006-82d3-66674cbd0e61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 842.379791] env[65522]: WARNING neutronclient.v2_0.client [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 842.380240] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.380395] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 842.380733] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 842.381616] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d98714c7-49c7-4f71-be90-b4ba6b65e28a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.391891] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Waiting for the task: (returnval){ [ 842.391891] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525ab524-06de-f65d-ae84-3b876700b9f6" [ 842.391891] env[65522]: _type = "Task" [ 842.391891] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.393539] env[65522]: WARNING openstack [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 842.393885] env[65522]: WARNING openstack [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 842.416976] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525ab524-06de-f65d-ae84-3b876700b9f6, 'name': SearchDatastore_Task, 'duration_secs': 0.011686} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.417319] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 842.417743] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 842.417837] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.417970] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 842.418865] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 842.418865] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13bcacf5-1125-4efc-afc4-252f504cbb8e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.428872] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 842.429108] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 842.429951] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f779381a-16ed-4f00-aea6-074dd041cddd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.441774] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Waiting for the task: (returnval){ [ 842.441774] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523f157e-1521-daf6-478d-cdbacae8fe38" [ 842.441774] env[65522]: _type = "Task" [ 842.441774] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.457562] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523f157e-1521-daf6-478d-cdbacae8fe38, 'name': SearchDatastore_Task, 'duration_secs': 0.011622} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.461395] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7311854-aba9-469e-a803-c99099cb8570 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.467677] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Waiting for the task: (returnval){ [ 842.467677] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529c0030-9f5c-46e2-c605-6b72426a1236" [ 842.467677] env[65522]: _type = "Task" [ 842.467677] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.477196] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529c0030-9f5c-46e2-c605-6b72426a1236, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.478804] env[65522]: WARNING neutronclient.v2_0.client [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 842.479447] env[65522]: WARNING openstack [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 842.479911] env[65522]: WARNING openstack [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 842.545913] env[65522]: WARNING openstack [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 842.545913] env[65522]: WARNING openstack [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 842.582183] env[65522]: DEBUG nova.network.neutron [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating instance_info_cache with network_info: [{"id": "9446c1bb-b443-4a52-879d-5fa6cb83f0dc", "address": "fa:16:3e:54:30:7c", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9446c1bb-b4", "ovs_interfaceid": "9446c1bb-b443-4a52-879d-5fa6cb83f0dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 842.608689] env[65522]: DEBUG nova.network.neutron [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 842.632383] env[65522]: WARNING openstack [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 842.632878] env[65522]: WARNING openstack [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 842.710685] env[65522]: WARNING neutronclient.v2_0.client [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 842.711361] env[65522]: WARNING openstack [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 842.711700] env[65522]: WARNING openstack [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 842.823160] env[65522]: DEBUG nova.network.neutron [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Updating instance_info_cache with network_info: [{"id": "cb206964-4331-495a-94ce-c2b95dccc600", "address": "fa:16:3e:5b:a0:80", "network": {"id": "922eeeeb-e124-4ae5-97a4-0e995b16f7ec", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-437359683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37fc01876567476f9b93d765b2cfddc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccf76700-491b-4462-ab19-e6d3a9ff87ac", "external-id": "nsx-vlan-transportzone-956", "segmentation_id": 956, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb206964-43", "ovs_interfaceid": "cb206964-4331-495a-94ce-c2b95dccc600", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 842.875077] env[65522]: DEBUG nova.compute.manager [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 842.883805] env[65522]: DEBUG oslo_concurrency.lockutils [req-1a762e38-b7c8-45a3-8249-c8ac12c78249 req-ac067d84-6ccc-4f38-82cd-13af32cb5a07 service nova] Releasing lock "refresh_cache-880390f5-00f1-4dea-9345-29c06289b340" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 842.897354] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f3fdb0-2aaa-4f1d-8f0a-788bb0a92a87 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.908146] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b040f4-ed88-4c5b-8754-893707518a34 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.940593] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee405f9d-9e1d-4abd-9f97-9629b1ecb414 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.950875] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa2b0e3-4c64-4833-982f-02004144220f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.969734] env[65522]: DEBUG nova.compute.provider_tree [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.982157] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529c0030-9f5c-46e2-c605-6b72426a1236, 'name': SearchDatastore_Task, 'duration_secs': 0.011304} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.983779] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 842.983779] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 880390f5-00f1-4dea-9345-29c06289b340/880390f5-00f1-4dea-9345-29c06289b340.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 842.983779] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b3160ea-66fe-4677-9f1a-44690a75b0a3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.994072] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Waiting for the task: (returnval){ [ 842.994072] env[65522]: value = "task-5114094" [ 842.994072] env[65522]: _type = "Task" [ 842.994072] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.004415] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114094, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.087123] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 843.327023] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Releasing lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 843.327564] env[65522]: DEBUG nova.compute.manager [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Instance network_info: |[{"id": "cb206964-4331-495a-94ce-c2b95dccc600", "address": "fa:16:3e:5b:a0:80", "network": {"id": "922eeeeb-e124-4ae5-97a4-0e995b16f7ec", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-437359683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37fc01876567476f9b93d765b2cfddc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccf76700-491b-4462-ab19-e6d3a9ff87ac", "external-id": "nsx-vlan-transportzone-956", "segmentation_id": 956, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb206964-43", "ovs_interfaceid": "cb206964-4331-495a-94ce-c2b95dccc600", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 843.328551] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:a0:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccf76700-491b-4462-ab19-e6d3a9ff87ac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cb206964-4331-495a-94ce-c2b95dccc600', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.341027] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 843.342159] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 843.342461] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32792041-c950-46b0-8cbe-ee8fed00b545 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.370502] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.370502] env[65522]: value = "task-5114095" [ 843.370502] env[65522]: _type = "Task" [ 843.370502] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.389720] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114095, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.407327] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 843.473993] env[65522]: DEBUG nova.compute.manager [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Received event network-vif-plugged-cb206964-4331-495a-94ce-c2b95dccc600 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 843.474310] env[65522]: DEBUG oslo_concurrency.lockutils [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Acquiring lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 843.474533] env[65522]: DEBUG oslo_concurrency.lockutils [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 843.474739] env[65522]: DEBUG oslo_concurrency.lockutils [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.474937] env[65522]: DEBUG nova.compute.manager [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] No waiting events found dispatching network-vif-plugged-cb206964-4331-495a-94ce-c2b95dccc600 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 843.475109] env[65522]: WARNING nova.compute.manager [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Received unexpected event network-vif-plugged-cb206964-4331-495a-94ce-c2b95dccc600 for instance with vm_state building and task_state spawning. [ 843.475280] env[65522]: DEBUG nova.compute.manager [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Received event network-changed-cb206964-4331-495a-94ce-c2b95dccc600 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 843.475440] env[65522]: DEBUG nova.compute.manager [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Refreshing instance network info cache due to event network-changed-cb206964-4331-495a-94ce-c2b95dccc600. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 843.475615] env[65522]: DEBUG oslo_concurrency.lockutils [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Acquiring lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.475746] env[65522]: DEBUG oslo_concurrency.lockutils [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Acquired lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 843.476113] env[65522]: DEBUG nova.network.neutron [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Refreshing network info cache for port cb206964-4331-495a-94ce-c2b95dccc600 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 843.478667] env[65522]: DEBUG nova.scheduler.client.report [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 843.506868] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114094, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.579029] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "1e7b8237-34ea-479e-b5a7-b1846661d61d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 843.579389] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "1e7b8237-34ea-479e-b5a7-b1846661d61d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 843.797047] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Acquiring lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 843.797369] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 843.797628] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Acquiring lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 843.797830] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 843.798086] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.801431] env[65522]: INFO nova.compute.manager [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Terminating instance [ 843.882399] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114095, 'name': CreateVM_Task, 'duration_secs': 0.457136} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.882583] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.883043] env[65522]: WARNING neutronclient.v2_0.client [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 843.883416] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.883578] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 843.883892] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 843.884204] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a789c25-82e9-45ae-a743-741dd400af84 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.890061] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 843.890061] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bebaa8-7996-dedb-56c0-5605aeb20c63" [ 843.890061] env[65522]: _type = "Task" [ 843.890061] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.898775] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bebaa8-7996-dedb-56c0-5605aeb20c63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.984393] env[65522]: WARNING neutronclient.v2_0.client [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 843.985102] env[65522]: WARNING openstack [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 843.985455] env[65522]: WARNING openstack [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 843.993192] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.862s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.993704] env[65522]: DEBUG nova.compute.manager [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 843.997317] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 38.245s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 843.997569] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.999708] env[65522]: DEBUG oslo_concurrency.lockutils [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 37.139s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 843.999891] env[65522]: DEBUG oslo_concurrency.lockutils [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.001787] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.851s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 844.001976] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.003692] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.780s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 844.005610] env[65522]: INFO nova.compute.claims [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 844.025507] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114094, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531377} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.025507] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 880390f5-00f1-4dea-9345-29c06289b340/880390f5-00f1-4dea-9345-29c06289b340.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 844.025806] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 844.025950] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c762308b-9f15-46c0-9032-46667ce2f9ed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.038473] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Waiting for the task: (returnval){ [ 844.038473] env[65522]: value = "task-5114096" [ 844.038473] env[65522]: _type = "Task" [ 844.038473] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.051683] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114096, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.058660] env[65522]: INFO nova.scheduler.client.report [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Deleted allocations for instance 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824 [ 844.061775] env[65522]: INFO nova.scheduler.client.report [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Deleted allocations for instance cd82f2fd-d657-4ece-8e90-9a47072f5546 [ 844.079421] env[65522]: INFO nova.scheduler.client.report [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Deleted allocations for instance 028c7fc5-5624-4dd5-9e2f-48191a86f765 [ 844.083603] env[65522]: DEBUG nova.compute.manager [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 844.137582] env[65522]: WARNING openstack [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 844.137582] env[65522]: WARNING openstack [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 844.210666] env[65522]: WARNING neutronclient.v2_0.client [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 844.211451] env[65522]: WARNING openstack [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 844.211813] env[65522]: WARNING openstack [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 844.299390] env[65522]: DEBUG nova.network.neutron [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Updated VIF entry in instance network info cache for port cb206964-4331-495a-94ce-c2b95dccc600. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 844.299802] env[65522]: DEBUG nova.network.neutron [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Updating instance_info_cache with network_info: [{"id": "cb206964-4331-495a-94ce-c2b95dccc600", "address": "fa:16:3e:5b:a0:80", "network": {"id": "922eeeeb-e124-4ae5-97a4-0e995b16f7ec", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-437359683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37fc01876567476f9b93d765b2cfddc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccf76700-491b-4462-ab19-e6d3a9ff87ac", "external-id": "nsx-vlan-transportzone-956", "segmentation_id": 956, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb206964-43", "ovs_interfaceid": "cb206964-4331-495a-94ce-c2b95dccc600", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 844.305602] env[65522]: DEBUG nova.compute.manager [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 844.305817] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 844.306745] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55315f92-9d4e-4dff-a1a1-c791ac59d5d1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.316180] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 844.317717] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-060ba54c-f3c0-4aab-a5ce-051a3b3a5a96 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.324528] env[65522]: DEBUG oslo_vmware.api [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Waiting for the task: (returnval){ [ 844.324528] env[65522]: value = "task-5114097" [ 844.324528] env[65522]: _type = "Task" [ 844.324528] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.333708] env[65522]: DEBUG oslo_vmware.api [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5114097, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.403116] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bebaa8-7996-dedb-56c0-5605aeb20c63, 'name': SearchDatastore_Task, 'duration_secs': 0.010956} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.403448] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 844.403680] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.403916] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.404073] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 844.404271] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.404556] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3cf7714a-6db8-4c26-8087-239124e41e0c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.414883] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.415109] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 844.415884] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc5c322c-1730-4a39-b145-db2e7a9446d7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.421694] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 844.421694] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52faef74-d5b3-7378-4335-f409611e5b94" [ 844.421694] env[65522]: _type = "Task" [ 844.421694] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.430972] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52faef74-d5b3-7378-4335-f409611e5b94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.514234] env[65522]: DEBUG nova.compute.utils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 844.515994] env[65522]: DEBUG nova.compute.manager [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 844.516219] env[65522]: DEBUG nova.network.neutron [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 844.516773] env[65522]: WARNING neutronclient.v2_0.client [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 844.516913] env[65522]: WARNING neutronclient.v2_0.client [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 844.517436] env[65522]: WARNING openstack [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 844.517842] env[65522]: WARNING openstack [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 844.548531] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114096, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.172498} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.548949] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 844.549855] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc531be8-0f53-4a45-803a-49a26e3a25d2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.574021] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 880390f5-00f1-4dea-9345-29c06289b340/880390f5-00f1-4dea-9345-29c06289b340.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 844.577787] env[65522]: DEBUG nova.policy [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f010fe7784043aaa3d4d44b921c438f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91597233ae9c44c094f4c32d90332fa6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 844.585740] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e248422-869f-412c-b935-2d10dfb2a351 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.601521] env[65522]: DEBUG oslo_concurrency.lockutils [None req-177070b0-ea20-45ee-a0ad-328fea2aebe4 tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.415s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.603095] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88c1400f-3c34-4763-b26a-f0e2dac835d0 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "cd82f2fd-d657-4ece-8e90-9a47072f5546" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.426s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.610414] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e240e51-c7c6-444f-bdb7-907d8dada30f tempest-ListServersNegativeTestJSON-1690492374 tempest-ListServersNegativeTestJSON-1690492374-project-member] Lock "028c7fc5-5624-4dd5-9e2f-48191a86f765" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.022s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.621798] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Waiting for the task: (returnval){ [ 844.621798] env[65522]: value = "task-5114098" [ 844.621798] env[65522]: _type = "Task" [ 844.621798] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.632612] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 844.632612] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7daf3985-e1d9-48f8-ad92-0a035a2fda2f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.637388] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114098, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.655439] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating instance '495fa98f-a9db-4214-87cc-d29209d3cb62' progress to 0 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 844.802641] env[65522]: DEBUG oslo_concurrency.lockutils [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Releasing lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 844.802917] env[65522]: DEBUG nova.compute.manager [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Received event network-changed-990aae71-9e8e-4da6-9e57-5c9257d127a4 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 844.803133] env[65522]: DEBUG nova.compute.manager [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Refreshing instance network info cache due to event network-changed-990aae71-9e8e-4da6-9e57-5c9257d127a4. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 844.803505] env[65522]: DEBUG oslo_concurrency.lockutils [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Acquiring lock "refresh_cache-18568baa-83de-4b54-addf-dee66211c8d7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.803505] env[65522]: DEBUG oslo_concurrency.lockutils [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Acquired lock "refresh_cache-18568baa-83de-4b54-addf-dee66211c8d7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 844.803773] env[65522]: DEBUG nova.network.neutron [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Refreshing network info cache for port 990aae71-9e8e-4da6-9e57-5c9257d127a4 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 844.835102] env[65522]: DEBUG oslo_vmware.api [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5114097, 'name': PowerOffVM_Task, 'duration_secs': 0.309026} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.835399] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 844.835568] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 844.835875] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ffe4b18e-dd9f-4d91-b8a2-2d3730b939c2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.847615] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "97dc8cac-8241-4912-a3ed-689439ef7ff8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 844.847885] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "97dc8cac-8241-4912-a3ed-689439ef7ff8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 844.848100] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "97dc8cac-8241-4912-a3ed-689439ef7ff8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 844.848292] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "97dc8cac-8241-4912-a3ed-689439ef7ff8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 844.848456] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "97dc8cac-8241-4912-a3ed-689439ef7ff8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.851054] env[65522]: INFO nova.compute.manager [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Terminating instance [ 844.902499] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 844.902688] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 844.903106] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Deleting the datastore file [datastore1] 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 844.903494] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-15e3494a-8931-44b2-9526-eb6687ccc07f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.912361] env[65522]: DEBUG oslo_vmware.api [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Waiting for the task: (returnval){ [ 844.912361] env[65522]: value = "task-5114100" [ 844.912361] env[65522]: _type = "Task" [ 844.912361] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.924213] env[65522]: DEBUG oslo_vmware.api [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5114100, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.936392] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52faef74-d5b3-7378-4335-f409611e5b94, 'name': SearchDatastore_Task, 'duration_secs': 0.016802} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.937522] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0291c231-6cfe-4f6b-ba8a-7010d6e0bef8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.945392] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 844.945392] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5256e9e9-edef-cc45-5483-d05ce27bad9e" [ 844.945392] env[65522]: _type = "Task" [ 844.945392] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.955304] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5256e9e9-edef-cc45-5483-d05ce27bad9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.020856] env[65522]: DEBUG nova.network.neutron [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Successfully created port: dc729a30-fef3-4b2e-ab41-6c7a4eb89f73 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 845.026578] env[65522]: DEBUG nova.compute.manager [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 845.133881] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.162697] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.162989] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c73a970c-a6b4-4739-9516-046720d88162 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.172837] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 845.172837] env[65522]: value = "task-5114101" [ 845.172837] env[65522]: _type = "Task" [ 845.172837] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.185089] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114101, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.307368] env[65522]: WARNING neutronclient.v2_0.client [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 845.309120] env[65522]: WARNING openstack [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 845.309228] env[65522]: WARNING openstack [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 845.355366] env[65522]: DEBUG nova.compute.manager [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 845.355577] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 845.356593] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a6c1e0-f24a-4f3d-bf4a-9756b35ab110 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.369183] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.371351] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33593788-d4ab-4166-be59-9a2c2557bfcb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.378513] env[65522]: DEBUG oslo_vmware.api [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 845.378513] env[65522]: value = "task-5114102" [ 845.378513] env[65522]: _type = "Task" [ 845.378513] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.390609] env[65522]: DEBUG oslo_vmware.api [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5114102, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.428814] env[65522]: DEBUG oslo_vmware.api [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Task: {'id': task-5114100, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183504} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.429642] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 845.430774] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 845.430774] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 845.430774] env[65522]: INFO nova.compute.manager [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Took 1.12 seconds to destroy the instance on the hypervisor. [ 845.430774] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 845.430774] env[65522]: DEBUG nova.compute.manager [-] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 845.431189] env[65522]: DEBUG nova.network.neutron [-] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 845.431189] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 845.435024] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 845.435024] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 845.458970] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5256e9e9-edef-cc45-5483-d05ce27bad9e, 'name': SearchDatastore_Task, 'duration_secs': 0.011488} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.463974] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 845.463974] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 77b3ca48-d2c6-4743-a44a-53b9aa84662a/77b3ca48-d2c6-4743-a44a-53b9aa84662a.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 845.463974] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7391e03d-c900-4591-93d8-333f097bd23b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.472496] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 845.472496] env[65522]: value = "task-5114103" [ 845.472496] env[65522]: _type = "Task" [ 845.472496] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.487221] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114103, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.630041] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41bc4802-8166-4142-9ab8-21e0c07bc74d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.638669] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114098, 'name': ReconfigVM_Task, 'duration_secs': 0.630743} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.640814] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 880390f5-00f1-4dea-9345-29c06289b340/880390f5-00f1-4dea-9345-29c06289b340.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 845.641769] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-184b9c5d-58fb-4ba9-aa96-d73feb10c89f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.643980] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67bd20c7-178f-4821-badc-183de702e8b2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.678452] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Waiting for the task: (returnval){ [ 845.678452] env[65522]: value = "task-5114104" [ 845.678452] env[65522]: _type = "Task" [ 845.678452] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.682727] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb3c1af-9989-404c-b2c6-6227f0eb1bd1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.694761] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114101, 'name': PowerOffVM_Task, 'duration_secs': 0.248822} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.699773] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 845.699947] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating instance '495fa98f-a9db-4214-87cc-d29209d3cb62' progress to 17 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 845.703398] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114104, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.704739] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf55f757-2a76-428a-9613-88b6a78bccba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.720796] env[65522]: DEBUG nova.compute.provider_tree [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.893265] env[65522]: DEBUG oslo_vmware.api [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5114102, 'name': PowerOffVM_Task, 'duration_secs': 0.251495} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.893733] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 845.894010] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 845.894399] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-276df914-9153-48f2-aab2-aa2e246fc3e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.897735] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 845.971385] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 845.971684] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 845.971908] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Deleting the datastore file [datastore2] 97dc8cac-8241-4912-a3ed-689439ef7ff8 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 845.972307] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75b5818c-a174-451e-b05b-ee257380b3d6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.984738] env[65522]: DEBUG oslo_vmware.api [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 845.984738] env[65522]: value = "task-5114106" [ 845.984738] env[65522]: _type = "Task" [ 845.984738] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.993366] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114103, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.001263] env[65522]: DEBUG oslo_vmware.api [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5114106, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.040718] env[65522]: DEBUG nova.compute.manager [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 846.079957] env[65522]: DEBUG nova.virt.hardware [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 846.079957] env[65522]: DEBUG nova.virt.hardware [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 846.079957] env[65522]: DEBUG nova.virt.hardware [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 846.079957] env[65522]: DEBUG nova.virt.hardware [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 846.080350] env[65522]: DEBUG nova.virt.hardware [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 846.080414] env[65522]: DEBUG nova.virt.hardware [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 846.080774] env[65522]: DEBUG nova.virt.hardware [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 846.081528] env[65522]: DEBUG nova.virt.hardware [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 846.081528] env[65522]: DEBUG nova.virt.hardware [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 846.081667] env[65522]: DEBUG nova.virt.hardware [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 846.081779] env[65522]: DEBUG nova.virt.hardware [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 846.087224] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b3edc1a-fc24-4a3e-95ae-51ad22d9f104 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.105635] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2adc6aa0-1414-4689-8bca-19f12c7f82dc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.189931] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114104, 'name': Rename_Task, 'duration_secs': 0.171993} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.189931] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 846.189931] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43e5affb-7c9f-4eb3-b636-6e04fe918c77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.197891] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Waiting for the task: (returnval){ [ 846.197891] env[65522]: value = "task-5114107" [ 846.197891] env[65522]: _type = "Task" [ 846.197891] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.207674] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114107, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.212828] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 846.213266] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 846.213504] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 846.213763] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 846.213968] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 846.214222] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 846.214546] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 846.214770] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 846.215035] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 846.215231] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 846.215428] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 846.221366] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8f0fc6e-0498-4cc2-8506-612c6939a8a9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.233107] env[65522]: DEBUG nova.scheduler.client.report [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 846.244172] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 846.244172] env[65522]: value = "task-5114108" [ 846.244172] env[65522]: _type = "Task" [ 846.244172] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.259122] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114108, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.485866] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114103, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.678608} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.489183] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 77b3ca48-d2c6-4743-a44a-53b9aa84662a/77b3ca48-d2c6-4743-a44a-53b9aa84662a.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.489404] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.489716] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-978f2931-69eb-49af-b90a-65705279a278 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.499849] env[65522]: DEBUG oslo_vmware.api [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5114106, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.311836} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.501231] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.501430] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 846.501609] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 846.501787] env[65522]: INFO nova.compute.manager [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 846.502092] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 846.502366] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 846.502366] env[65522]: value = "task-5114109" [ 846.502366] env[65522]: _type = "Task" [ 846.502366] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.502559] env[65522]: DEBUG nova.compute.manager [-] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 846.502674] env[65522]: DEBUG nova.network.neutron [-] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 846.502902] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.503452] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.503721] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.521570] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114109, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.662815] env[65522]: DEBUG nova.network.neutron [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Successfully updated port: dc729a30-fef3-4b2e-ab41-6c7a4eb89f73 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 846.707829] env[65522]: WARNING openstack [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.708324] env[65522]: WARNING openstack [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.723027] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114107, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.738763] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.735s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 846.740962] env[65522]: DEBUG nova.compute.manager [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 846.744034] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 36.693s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 846.746140] env[65522]: INFO nova.compute.claims [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.763173] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114108, 'name': ReconfigVM_Task, 'duration_secs': 0.381637} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.763492] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating instance '495fa98f-a9db-4214-87cc-d29209d3cb62' progress to 33 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 846.790909] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.933864] env[65522]: WARNING neutronclient.v2_0.client [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.933864] env[65522]: WARNING openstack [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.933864] env[65522]: WARNING openstack [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.032074] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114109, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078327} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.032555] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.033563] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-952cabac-3e5d-4240-bc71-da24d61e1ab0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.060797] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] 77b3ca48-d2c6-4743-a44a-53b9aa84662a/77b3ca48-d2c6-4743-a44a-53b9aa84662a.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.061155] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbbf71fa-5791-4317-b143-789fbe1ea695 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.076684] env[65522]: DEBUG nova.network.neutron [-] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 847.086035] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 847.086035] env[65522]: value = "task-5114110" [ 847.086035] env[65522]: _type = "Task" [ 847.086035] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.095919] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114110, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.136725] env[65522]: DEBUG nova.network.neutron [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Updated VIF entry in instance network info cache for port 990aae71-9e8e-4da6-9e57-5c9257d127a4. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 847.137312] env[65522]: DEBUG nova.network.neutron [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Updating instance_info_cache with network_info: [{"id": "990aae71-9e8e-4da6-9e57-5c9257d127a4", "address": "fa:16:3e:1b:6a:ed", "network": {"id": "5165f03c-e790-4121-b5d7-3c5176afdd7e", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-551605283-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd6e7f7b6a224cac8b8e52bf19032911", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "edd47158-6f4b-44a1-8e82-0411205ad299", "external-id": "nsx-vlan-transportzone-587", "segmentation_id": 587, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap990aae71-9e", "ovs_interfaceid": "990aae71-9e8e-4da6-9e57-5c9257d127a4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 847.166189] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "refresh_cache-7617189c-a902-42e7-8165-0e7c4a1de06d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.166373] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "refresh_cache-7617189c-a902-42e7-8165-0e7c4a1de06d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 847.167739] env[65522]: DEBUG nova.network.neutron [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 847.211888] env[65522]: DEBUG oslo_vmware.api [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114107, 'name': PowerOnVM_Task, 'duration_secs': 0.98877} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.212561] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 847.212778] env[65522]: INFO nova.compute.manager [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Took 9.72 seconds to spawn the instance on the hypervisor. [ 847.212967] env[65522]: DEBUG nova.compute.manager [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 847.213819] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec28a5fe-511b-4f9b-b040-a18babdd7321 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.252791] env[65522]: DEBUG nova.compute.utils [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 847.253984] env[65522]: DEBUG nova.compute.manager [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Not allocating networking since 'none' was specified. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 847.271956] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 847.271956] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 847.271956] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 847.272909] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 847.272909] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 847.272909] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 847.272909] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.272909] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 847.273515] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 847.273515] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 847.273965] env[65522]: DEBUG nova.virt.hardware [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 847.279504] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Reconfiguring VM instance instance-0000002b to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 847.282771] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c77db7b2-62c0-4fd0-a012-ee2b041457af {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.307763] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 847.307763] env[65522]: value = "task-5114111" [ 847.307763] env[65522]: _type = "Task" [ 847.307763] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.320351] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114111, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.452758] env[65522]: DEBUG nova.compute.manager [req-e40fce89-449c-470b-b68c-311c57c9ad1a req-bd0597fe-61d7-457c-b94b-69191dc1d478 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Received event network-vif-plugged-dc729a30-fef3-4b2e-ab41-6c7a4eb89f73 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 847.452986] env[65522]: DEBUG oslo_concurrency.lockutils [req-e40fce89-449c-470b-b68c-311c57c9ad1a req-bd0597fe-61d7-457c-b94b-69191dc1d478 service nova] Acquiring lock "7617189c-a902-42e7-8165-0e7c4a1de06d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 847.453216] env[65522]: DEBUG oslo_concurrency.lockutils [req-e40fce89-449c-470b-b68c-311c57c9ad1a req-bd0597fe-61d7-457c-b94b-69191dc1d478 service nova] Lock "7617189c-a902-42e7-8165-0e7c4a1de06d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 847.453387] env[65522]: DEBUG oslo_concurrency.lockutils [req-e40fce89-449c-470b-b68c-311c57c9ad1a req-bd0597fe-61d7-457c-b94b-69191dc1d478 service nova] Lock "7617189c-a902-42e7-8165-0e7c4a1de06d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 847.453555] env[65522]: DEBUG nova.compute.manager [req-e40fce89-449c-470b-b68c-311c57c9ad1a req-bd0597fe-61d7-457c-b94b-69191dc1d478 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] No waiting events found dispatching network-vif-plugged-dc729a30-fef3-4b2e-ab41-6c7a4eb89f73 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 847.453713] env[65522]: WARNING nova.compute.manager [req-e40fce89-449c-470b-b68c-311c57c9ad1a req-bd0597fe-61d7-457c-b94b-69191dc1d478 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Received unexpected event network-vif-plugged-dc729a30-fef3-4b2e-ab41-6c7a4eb89f73 for instance with vm_state building and task_state spawning. [ 847.472123] env[65522]: DEBUG nova.compute.manager [req-305d01ad-4dfa-4b19-ad70-8c2a75d7a62a req-21d528e6-5deb-4f62-948c-2f1dc548985f service nova] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Received event network-vif-deleted-ef361904-5296-49e1-aeb2-182cd4adb91b {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 847.579832] env[65522]: INFO nova.compute.manager [-] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Took 2.15 seconds to deallocate network for instance. [ 847.602728] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114110, 'name': ReconfigVM_Task, 'duration_secs': 0.400182} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.602952] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Reconfigured VM instance instance-00000036 to attach disk [datastore2] 77b3ca48-d2c6-4743-a44a-53b9aa84662a/77b3ca48-d2c6-4743-a44a-53b9aa84662a.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.603688] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1bbc35b0-653a-41d9-a35c-844b2b14dbd6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.611410] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 847.611410] env[65522]: value = "task-5114112" [ 847.611410] env[65522]: _type = "Task" [ 847.611410] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.621506] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114112, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.640737] env[65522]: DEBUG oslo_concurrency.lockutils [req-19104166-38ac-4d9b-9cd6-3b0328f4b91d req-680ef537-e59e-41b0-b36b-6aad73042d5b service nova] Releasing lock "refresh_cache-18568baa-83de-4b54-addf-dee66211c8d7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 847.659515] env[65522]: DEBUG nova.network.neutron [-] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 847.670867] env[65522]: WARNING openstack [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.671285] env[65522]: WARNING openstack [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.724681] env[65522]: DEBUG nova.network.neutron [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 847.733987] env[65522]: INFO nova.compute.manager [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Took 50.13 seconds to build instance. [ 847.749653] env[65522]: WARNING openstack [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.749653] env[65522]: WARNING openstack [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.758365] env[65522]: DEBUG nova.compute.manager [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 847.818420] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114111, 'name': ReconfigVM_Task, 'duration_secs': 0.401265} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.818540] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Reconfigured VM instance instance-0000002b to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 847.819360] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dee84c1-950d-4bf3-9df2-0af2993d939b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.848108] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] 495fa98f-a9db-4214-87cc-d29209d3cb62/495fa98f-a9db-4214-87cc-d29209d3cb62.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.851469] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-859ce2e0-a3cd-4b3e-aa90-6b6e88f035b5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.866430] env[65522]: WARNING neutronclient.v2_0.client [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 847.867233] env[65522]: WARNING openstack [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.867757] env[65522]: WARNING openstack [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.883636] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 847.883636] env[65522]: value = "task-5114113" [ 847.883636] env[65522]: _type = "Task" [ 847.883636] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.893476] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114113, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.977027] env[65522]: DEBUG nova.network.neutron [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Updating instance_info_cache with network_info: [{"id": "dc729a30-fef3-4b2e-ab41-6c7a4eb89f73", "address": "fa:16:3e:6a:3b:a6", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc729a30-fe", "ovs_interfaceid": "dc729a30-fef3-4b2e-ab41-6c7a4eb89f73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 848.097423] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 848.123719] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114112, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.150650] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.151074] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.162457] env[65522]: INFO nova.compute.manager [-] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Took 1.66 seconds to deallocate network for instance. [ 848.237807] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fd238a41-6ccb-43fe-be7d-b5613b87ec64 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Lock "880390f5-00f1-4dea-9345-29c06289b340" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.640s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 848.343634] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17eb3397-a6d6-41a8-9d25-9b9b716a1516 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.352856] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35eb92f4-205f-419d-af59-9af590382b64 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.390860] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-930f1f1c-b1ee-4971-bcc0-5c9286f348c9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.405285] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22f282e-db79-41af-8879-5f3adb7e1721 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.410378] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114113, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.422034] env[65522]: DEBUG nova.compute.provider_tree [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.480246] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "refresh_cache-7617189c-a902-42e7-8165-0e7c4a1de06d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 848.481044] env[65522]: DEBUG nova.compute.manager [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Instance network_info: |[{"id": "dc729a30-fef3-4b2e-ab41-6c7a4eb89f73", "address": "fa:16:3e:6a:3b:a6", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc729a30-fe", "ovs_interfaceid": "dc729a30-fef3-4b2e-ab41-6c7a4eb89f73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 848.481338] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:3b:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56136ef6-99d7-4562-9a9f-d66fec951c5c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dc729a30-fef3-4b2e-ab41-6c7a4eb89f73', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 848.489895] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Creating folder: Project (91597233ae9c44c094f4c32d90332fa6). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.490266] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b933a141-1c22-46ef-a50f-c29351c619ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.503337] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Created folder: Project (91597233ae9c44c094f4c32d90332fa6) in parent group-v994660. [ 848.503550] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Creating folder: Instances. Parent ref: group-v994811. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.503859] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3211a576-93b4-4d97-9d50-e03fccfad91f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.516105] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Created folder: Instances in parent group-v994811. [ 848.516295] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 848.516538] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 848.516842] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-43d222cb-2357-45b9-9340-3fe60052ea8a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.541808] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 848.541808] env[65522]: value = "task-5114116" [ 848.541808] env[65522]: _type = "Task" [ 848.541808] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.551888] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114116, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.623788] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114112, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.660877] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.661112] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.661273] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.661441] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.661665] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.661833] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.662408] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 848.662408] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.672559] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 848.776198] env[65522]: DEBUG nova.compute.manager [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 848.811237] env[65522]: DEBUG nova.virt.hardware [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 848.811598] env[65522]: DEBUG nova.virt.hardware [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 848.811822] env[65522]: DEBUG nova.virt.hardware [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 848.812149] env[65522]: DEBUG nova.virt.hardware [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 848.812353] env[65522]: DEBUG nova.virt.hardware [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 848.812590] env[65522]: DEBUG nova.virt.hardware [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 848.812834] env[65522]: DEBUG nova.virt.hardware [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 848.813064] env[65522]: DEBUG nova.virt.hardware [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 848.813328] env[65522]: DEBUG nova.virt.hardware [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 848.813524] env[65522]: DEBUG nova.virt.hardware [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 848.813763] env[65522]: DEBUG nova.virt.hardware [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 848.815340] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22ae47a-125b-4329-95d4-f5d2b6f8a5af {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.826383] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b48a676-850b-439d-abd8-2cd86d3e9dec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.843189] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 848.848953] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Creating folder: Project (78940f9e22cf4de2b41934ab3fa76abc). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.849390] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c7db0ec2-d82d-4eec-812b-b1867286ec42 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.862444] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Created folder: Project (78940f9e22cf4de2b41934ab3fa76abc) in parent group-v994660. [ 848.862786] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Creating folder: Instances. Parent ref: group-v994814. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.862916] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2b2563ce-d1f5-4829-8ba5-b29f510b9386 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.874586] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Created folder: Instances in parent group-v994814. [ 848.874901] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 848.875150] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 848.875400] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4c45c16d-cace-4e76-ad81-95bfbe43ed1c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.898419] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Acquiring lock "880390f5-00f1-4dea-9345-29c06289b340" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 848.898922] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Lock "880390f5-00f1-4dea-9345-29c06289b340" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 848.898922] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Acquiring lock "880390f5-00f1-4dea-9345-29c06289b340-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 848.899156] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Lock "880390f5-00f1-4dea-9345-29c06289b340-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 848.899230] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Lock "880390f5-00f1-4dea-9345-29c06289b340-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 848.900970] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114113, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.902542] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 848.902542] env[65522]: value = "task-5114119" [ 848.902542] env[65522]: _type = "Task" [ 848.902542] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.903175] env[65522]: INFO nova.compute.manager [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Terminating instance [ 848.916852] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114119, 'name': CreateVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.926373] env[65522]: DEBUG nova.scheduler.client.report [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 849.054791] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114116, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.125679] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114112, 'name': Rename_Task, 'duration_secs': 1.215629} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.125991] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 849.126282] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8307d53b-a7c2-451c-ba72-140bc2d20c51 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.133234] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 849.133234] env[65522]: value = "task-5114120" [ 849.133234] env[65522]: _type = "Task" [ 849.133234] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.144194] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114120, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.166719] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 849.399685] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114113, 'name': ReconfigVM_Task, 'duration_secs': 1.26825} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.399932] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Reconfigured VM instance instance-0000002b to attach disk [datastore1] 495fa98f-a9db-4214-87cc-d29209d3cb62/495fa98f-a9db-4214-87cc-d29209d3cb62.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.400248] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating instance '495fa98f-a9db-4214-87cc-d29209d3cb62' progress to 50 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 849.411995] env[65522]: DEBUG nova.compute.manager [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 849.411995] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 849.417075] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70be4484-b656-4bcb-9e79-62dbc87d3a42 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.419846] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114119, 'name': CreateVM_Task, 'duration_secs': 0.335192} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.420037] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 849.420856] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.421066] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 849.421380] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 849.421710] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b3520cc-ba7a-4ba0-bd99-e705daeec36d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.425838] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 849.426674] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1786c613-0186-48e9-bf23-d80a3e84d492 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.429740] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 849.429740] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527e8069-e0d7-0dda-7df8-019ff3c714a1" [ 849.429740] env[65522]: _type = "Task" [ 849.429740] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.430585] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.687s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.431043] env[65522]: DEBUG nova.compute.manager [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 849.444021] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.482s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.444021] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.444021] env[65522]: DEBUG oslo_concurrency.lockutils [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.720s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.444021] env[65522]: DEBUG oslo_concurrency.lockutils [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.445922] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.532s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.446160] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.448141] env[65522]: DEBUG oslo_concurrency.lockutils [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.258s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.448330] env[65522]: DEBUG oslo_concurrency.lockutils [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.450722] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.256s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.452018] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.452953] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.924s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.454683] env[65522]: INFO nova.compute.claims [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 849.458718] env[65522]: DEBUG oslo_vmware.api [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Waiting for the task: (returnval){ [ 849.458718] env[65522]: value = "task-5114121" [ 849.458718] env[65522]: _type = "Task" [ 849.458718] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.469405] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527e8069-e0d7-0dda-7df8-019ff3c714a1, 'name': SearchDatastore_Task, 'duration_secs': 0.011419} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.471801] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 849.472422] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 849.472606] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.472905] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 849.473314] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 849.477997] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f3bfe55-0b8d-4f61-a36c-9c6c5caac5ac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.480726] env[65522]: DEBUG oslo_vmware.api [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114121, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.489565] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 849.489879] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 849.490799] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8356ccf7-bb54-42e0-b88f-5283f630a42c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.495364] env[65522]: INFO nova.scheduler.client.report [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Deleted allocations for instance a8f9f223-023a-48fc-9186-a360a78e27a9 [ 849.496718] env[65522]: INFO nova.scheduler.client.report [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Deleted allocations for instance 9fa6e208-f913-4ecc-8548-ed6503fe6273 [ 849.505311] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 849.505311] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526496da-6bba-da29-686c-db0b990f03bb" [ 849.505311] env[65522]: _type = "Task" [ 849.505311] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.516120] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526496da-6bba-da29-686c-db0b990f03bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.550629] env[65522]: INFO nova.scheduler.client.report [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Deleted allocations for instance 7693f017-69b3-4ec4-ae3f-7e65ed34a471 [ 849.553205] env[65522]: INFO nova.scheduler.client.report [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Deleted allocations for instance cdee3a32-ad51-4364-8440-4744e8df322f [ 849.559106] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114116, 'name': CreateVM_Task, 'duration_secs': 0.516206} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.565619] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 849.566690] env[65522]: WARNING neutronclient.v2_0.client [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 849.567433] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.568236] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 849.568236] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 849.568384] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fa4c076-9a32-4a34-8ad3-3522e4234365 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.574940] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 849.574940] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f52d62-24a7-ceeb-6086-c5cc08d18533" [ 849.574940] env[65522]: _type = "Task" [ 849.574940] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.584995] env[65522]: DEBUG nova.compute.manager [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Received event network-changed-dc729a30-fef3-4b2e-ab41-6c7a4eb89f73 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 849.585848] env[65522]: DEBUG nova.compute.manager [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Refreshing instance network info cache due to event network-changed-dc729a30-fef3-4b2e-ab41-6c7a4eb89f73. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 849.585848] env[65522]: DEBUG oslo_concurrency.lockutils [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] Acquiring lock "refresh_cache-7617189c-a902-42e7-8165-0e7c4a1de06d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.585848] env[65522]: DEBUG oslo_concurrency.lockutils [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] Acquired lock "refresh_cache-7617189c-a902-42e7-8165-0e7c4a1de06d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 849.585848] env[65522]: DEBUG nova.network.neutron [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Refreshing network info cache for port dc729a30-fef3-4b2e-ab41-6c7a4eb89f73 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 849.590884] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f52d62-24a7-ceeb-6086-c5cc08d18533, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.594650] env[65522]: INFO nova.scheduler.client.report [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Deleted allocations for instance d2fdffb8-76d1-4b74-819f-25eb29ca91ce [ 849.612658] env[65522]: DEBUG nova.compute.manager [req-aa2bbee5-0641-42c1-bfac-dad2e5eb38f6 req-5c33434b-1a9c-40cd-a476-2babd9b52e41 service nova] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Received event network-vif-deleted-475e049b-993f-4ab4-aba6-da16d2828c40 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 849.647857] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114120, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.907125] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79929766-b5d7-4b9f-9b49-1a67ac26a9e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.929998] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c224ab9d-f968-4367-bdaa-ad16b932fe07 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.948738] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating instance '495fa98f-a9db-4214-87cc-d29209d3cb62' progress to 67 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 849.961169] env[65522]: DEBUG nova.compute.utils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 849.965258] env[65522]: DEBUG nova.compute.manager [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 849.965258] env[65522]: DEBUG nova.network.neutron [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 849.966024] env[65522]: WARNING neutronclient.v2_0.client [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 849.966134] env[65522]: WARNING neutronclient.v2_0.client [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 849.966639] env[65522]: WARNING openstack [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 849.966974] env[65522]: WARNING openstack [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 849.986022] env[65522]: DEBUG oslo_vmware.api [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114121, 'name': PowerOffVM_Task, 'duration_secs': 0.185006} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.986022] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 849.986022] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 849.986022] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-20bb890e-3f9f-4523-8a4e-eb4ae367626d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.005338] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7739ef29-75f9-4cdd-84ca-c58cf9472bf8 tempest-ServersAdminNegativeTestJSON-1015317740 tempest-ServersAdminNegativeTestJSON-1015317740-project-member] Lock "a8f9f223-023a-48fc-9186-a360a78e27a9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.811s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.014597] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c6c7b500-e254-4f98-8820-c381de6262c7 tempest-InstanceActionsTestJSON-594736579 tempest-InstanceActionsTestJSON-594736579-project-member] Lock "9fa6e208-f913-4ecc-8548-ed6503fe6273" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.354s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.024488] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526496da-6bba-da29-686c-db0b990f03bb, 'name': SearchDatastore_Task, 'duration_secs': 0.012807} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.025369] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b350d348-94c4-4a94-a23c-3024e61a6e8f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.034119] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 850.034119] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5284971b-e24f-eb97-2241-e2fb7a1398c0" [ 850.034119] env[65522]: _type = "Task" [ 850.034119] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.045744] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5284971b-e24f-eb97-2241-e2fb7a1398c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.053926] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 850.054457] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 850.054669] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Deleting the datastore file [datastore1] 880390f5-00f1-4dea-9345-29c06289b340 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 850.054959] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ea756ffb-f43d-455c-9318-799b68e617d2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.064315] env[65522]: DEBUG oslo_vmware.api [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Waiting for the task: (returnval){ [ 850.064315] env[65522]: value = "task-5114123" [ 850.064315] env[65522]: _type = "Task" [ 850.064315] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.072188] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7e8d4342-e2b1-4032-85a8-db3f8c9cecb4 tempest-InstanceActionsNegativeTestJSON-514668720 tempest-InstanceActionsNegativeTestJSON-514668720-project-member] Lock "cdee3a32-ad51-4364-8440-4744e8df322f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.737s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.074034] env[65522]: DEBUG oslo_concurrency.lockutils [None req-80dfdc72-49a1-4949-8a19-6b1beb1eb51c tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "7693f017-69b3-4ec4-ae3f-7e65ed34a471" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.965s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.079794] env[65522]: DEBUG oslo_vmware.api [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114123, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.091035] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f52d62-24a7-ceeb-6086-c5cc08d18533, 'name': SearchDatastore_Task, 'duration_secs': 0.012181} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.094794] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 850.094794] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.094794] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.094794] env[65522]: WARNING neutronclient.v2_0.client [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 850.094794] env[65522]: WARNING openstack [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 850.094794] env[65522]: WARNING openstack [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 850.106636] env[65522]: DEBUG oslo_concurrency.lockutils [None req-97573198-794f-4130-a2dc-b9f64476d554 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464 tempest-FloatingIPsAssociationNegativeTestJSON-1331977464-project-member] Lock "d2fdffb8-76d1-4b74-819f-25eb29ca91ce" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.378s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.146440] env[65522]: DEBUG oslo_vmware.api [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114120, 'name': PowerOnVM_Task, 'duration_secs': 0.565666} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.146753] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 850.146989] env[65522]: INFO nova.compute.manager [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Took 9.77 seconds to spawn the instance on the hypervisor. [ 850.147201] env[65522]: DEBUG nova.compute.manager [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 850.148367] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec213d3-e5e1-448f-871c-06767a28c31f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.457222] env[65522]: WARNING neutronclient.v2_0.client [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 850.483473] env[65522]: DEBUG nova.compute.manager [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 850.553329] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5284971b-e24f-eb97-2241-e2fb7a1398c0, 'name': SearchDatastore_Task, 'duration_secs': 0.010557} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.553781] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 850.554170] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 26432ad6-610e-49c2-8699-f1e7e128e5b8/26432ad6-610e-49c2-8699-f1e7e128e5b8.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 850.554730] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 850.555208] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.556026] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-229a1885-c3c0-4179-80af-86942aa51d87 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.559628] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fea78902-a653-4e7d-b1bb-a2a73d9e55df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.571691] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 850.571691] env[65522]: value = "task-5114124" [ 850.571691] env[65522]: _type = "Task" [ 850.571691] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.577225] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.577464] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 850.579398] env[65522]: DEBUG oslo_vmware.api [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Task: {'id': task-5114123, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192302} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.584054] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-862b2e22-f3a1-4e45-be40-23c9e17ed8b5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.586819] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 850.587050] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 850.587275] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 850.587511] env[65522]: INFO nova.compute.manager [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Took 1.18 seconds to destroy the instance on the hypervisor. [ 850.587844] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 850.588420] env[65522]: DEBUG nova.compute.manager [-] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 850.588503] env[65522]: DEBUG nova.network.neutron [-] [instance: 880390f5-00f1-4dea-9345-29c06289b340] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 850.588988] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 850.589692] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 850.590028] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 850.606703] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 850.606703] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fa885b-be18-c6c8-325b-1f1c5defb766" [ 850.606703] env[65522]: _type = "Task" [ 850.606703] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.611309] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114124, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.621241] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fa885b-be18-c6c8-325b-1f1c5defb766, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.668735] env[65522]: INFO nova.compute.manager [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Took 50.93 seconds to build instance. [ 850.744093] env[65522]: DEBUG nova.network.neutron [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Port 9446c1bb-b443-4a52-879d-5fa6cb83f0dc binding to destination host cpu-1 is already ACTIVE {{(pid=65522) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 850.866367] env[65522]: DEBUG nova.policy [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1103ba172d754007bb2b7339e1d674f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '503adb79a1f74f1b98e1f06ec54f1584', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 850.899316] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 851.091423] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114124, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.094384] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5600423e-8437-47ef-bc5c-2438058ffa1b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.105524] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a163bee9-9542-472f-b362-90cdfbfa6618 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.149654] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22b2ca3d-debe-4767-93ff-24abf111ca56 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.152400] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fa885b-be18-c6c8-325b-1f1c5defb766, 'name': SearchDatastore_Task, 'duration_secs': 0.018412} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.153755] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3da1809e-b7ee-49f8-86a5-a7fb54beee6d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.160407] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b08a62-02db-4789-a541-9d520e8f775b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.166402] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 851.166402] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d8ba22-9a4b-7d35-38f4-d45c9a81d224" [ 851.166402] env[65522]: _type = "Task" [ 851.166402] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.178869] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0deb5f2b-9b12-4233-9266-58846774d931 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.448s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 851.179502] env[65522]: DEBUG nova.compute.provider_tree [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.187769] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d8ba22-9a4b-7d35-38f4-d45c9a81d224, 'name': SearchDatastore_Task, 'duration_secs': 0.011115} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.188074] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 851.188339] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 7617189c-a902-42e7-8165-0e7c4a1de06d/7617189c-a902-42e7-8165-0e7c4a1de06d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 851.188615] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-805a3335-5a97-4ae8-95d0-4153e2322e8b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.196852] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 851.196852] env[65522]: value = "task-5114125" [ 851.196852] env[65522]: _type = "Task" [ 851.196852] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.206870] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114125, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.472713] env[65522]: DEBUG nova.network.neutron [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Successfully created port: 3f79042c-23dd-4fc0-b3cf-1e2884edabc1 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 851.492147] env[65522]: WARNING openstack [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 851.492568] env[65522]: WARNING openstack [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 851.504188] env[65522]: DEBUG nova.compute.manager [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 851.537878] env[65522]: DEBUG nova.virt.hardware [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:01:50Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1d8f2c10-9623-4a70-b964-018e648fd761',id=39,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-811894183',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 851.538325] env[65522]: DEBUG nova.virt.hardware [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 851.538588] env[65522]: DEBUG nova.virt.hardware [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 851.539680] env[65522]: DEBUG nova.virt.hardware [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 851.540171] env[65522]: DEBUG nova.virt.hardware [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 851.540469] env[65522]: DEBUG nova.virt.hardware [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 851.540846] env[65522]: DEBUG nova.virt.hardware [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.541173] env[65522]: DEBUG nova.virt.hardware [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 851.541447] env[65522]: DEBUG nova.virt.hardware [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 851.543260] env[65522]: DEBUG nova.virt.hardware [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 851.543260] env[65522]: DEBUG nova.virt.hardware [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 851.543571] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a628f067-1dc7-43b3-8e14-ad77c389d29f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.561358] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b338e3-c929-4c1d-8450-eb4fc2d5438f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.598406] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114124, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.591501} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.598633] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 26432ad6-610e-49c2-8699-f1e7e128e5b8/26432ad6-610e-49c2-8699-f1e7e128e5b8.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 851.598899] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 851.599235] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e93725ca-4b1d-42f6-982e-f63a4f283c35 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.611247] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 851.611247] env[65522]: value = "task-5114126" [ 851.611247] env[65522]: _type = "Task" [ 851.611247] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.627581] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114126, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.683726] env[65522]: DEBUG nova.scheduler.client.report [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 851.711570] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114125, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.499451} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.712983] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 7617189c-a902-42e7-8165-0e7c4a1de06d/7617189c-a902-42e7-8165-0e7c4a1de06d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 851.713358] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 851.713784] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c00b3570-3464-4ac1-ac5f-07ff5f24a540 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.722978] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 851.722978] env[65522]: value = "task-5114127" [ 851.722978] env[65522]: _type = "Task" [ 851.722978] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.736021] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114127, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.775629] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "495fa98f-a9db-4214-87cc-d29209d3cb62-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 851.775754] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "495fa98f-a9db-4214-87cc-d29209d3cb62-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 851.775880] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "495fa98f-a9db-4214-87cc-d29209d3cb62-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 852.118138] env[65522]: DEBUG nova.network.neutron [-] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 852.122526] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114126, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097818} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.123070] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 852.123867] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e37a1e6c-ef33-4eae-a494-cc6c8bd46089 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.145811] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 26432ad6-610e-49c2-8699-f1e7e128e5b8/26432ad6-610e-49c2-8699-f1e7e128e5b8.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.146897] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89ce731e-08e8-4e20-ac04-4303ce291e15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.170158] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 852.170158] env[65522]: value = "task-5114128" [ 852.170158] env[65522]: _type = "Task" [ 852.170158] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.180720] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114128, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.191861] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.739s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 852.192549] env[65522]: DEBUG nova.compute.manager [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 852.195528] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.953s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 852.195775] env[65522]: DEBUG nova.objects.instance [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Lazy-loading 'resources' on Instance uuid 1be995d0-3141-4dcd-ad9a-278d9874e457 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.234246] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114127, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.207098} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.234537] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 852.235452] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37097e6e-de74-4aa0-b47c-57f360b9cd1f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.262279] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 7617189c-a902-42e7-8165-0e7c4a1de06d/7617189c-a902-42e7-8165-0e7c4a1de06d.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.262279] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07d5007f-683d-4326-8764-b8900b36ad23 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.286745] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 852.286745] env[65522]: value = "task-5114129" [ 852.286745] env[65522]: _type = "Task" [ 852.286745] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.299804] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114129, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.624561] env[65522]: INFO nova.compute.manager [-] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Took 2.04 seconds to deallocate network for instance. [ 852.682623] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114128, 'name': ReconfigVM_Task, 'duration_secs': 0.319445} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.683054] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 26432ad6-610e-49c2-8699-f1e7e128e5b8/26432ad6-610e-49c2-8699-f1e7e128e5b8.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 852.683659] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d36a50ce-7cd3-4583-b71a-88983721f38c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.694231] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 852.694231] env[65522]: value = "task-5114130" [ 852.694231] env[65522]: _type = "Task" [ 852.694231] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.700812] env[65522]: DEBUG nova.compute.utils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 852.708446] env[65522]: DEBUG nova.compute.manager [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 852.709221] env[65522]: DEBUG nova.network.neutron [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 852.709221] env[65522]: WARNING neutronclient.v2_0.client [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 852.709559] env[65522]: WARNING neutronclient.v2_0.client [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 852.710183] env[65522]: WARNING openstack [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 852.710518] env[65522]: WARNING openstack [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 852.727085] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114130, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.781980] env[65522]: WARNING neutronclient.v2_0.client [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 852.805339] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114129, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.108997] env[65522]: WARNING neutronclient.v2_0.client [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 853.109744] env[65522]: WARNING openstack [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 853.110117] env[65522]: WARNING openstack [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 853.132193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 853.195825] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045edfcd-e962-4d5a-af78-70892a9b92ab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.210662] env[65522]: DEBUG nova.compute.manager [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 853.218058] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10feac7e-0bd8-4dee-bb4c-fd471f38fd08 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.221724] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114130, 'name': Rename_Task, 'duration_secs': 0.169555} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.222440] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 853.224018] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2680f555-86aa-4c5e-b332-4ba59c6b6eae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.259947] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c162515e-6682-43a6-b890-763a0a7961cb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.265375] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 853.265375] env[65522]: value = "task-5114131" [ 853.265375] env[65522]: _type = "Task" [ 853.265375] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.278128] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f8c20a-422a-4416-ae47-25f2d4202089 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.284939] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114131, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.301768] env[65522]: DEBUG nova.compute.provider_tree [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.306115] env[65522]: DEBUG nova.network.neutron [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Successfully updated port: 3f79042c-23dd-4fc0-b3cf-1e2884edabc1 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 853.311162] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114129, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.383520] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.383719] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 853.383912] env[65522]: DEBUG nova.network.neutron [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 853.574990] env[65522]: DEBUG nova.policy [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '60b53dc583e44c56b2465a0b45cbbd69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b5c2d36708f44bb9a0aa34f49014a3ed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 853.613145] env[65522]: DEBUG nova.network.neutron [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Updated VIF entry in instance network info cache for port dc729a30-fef3-4b2e-ab41-6c7a4eb89f73. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 853.613518] env[65522]: DEBUG nova.network.neutron [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Updating instance_info_cache with network_info: [{"id": "dc729a30-fef3-4b2e-ab41-6c7a4eb89f73", "address": "fa:16:3e:6a:3b:a6", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc729a30-fe", "ovs_interfaceid": "dc729a30-fef3-4b2e-ab41-6c7a4eb89f73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 853.779241] env[65522]: DEBUG oslo_vmware.api [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114131, 'name': PowerOnVM_Task, 'duration_secs': 0.482509} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.779572] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 853.779805] env[65522]: INFO nova.compute.manager [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Took 5.00 seconds to spawn the instance on the hypervisor. [ 853.780110] env[65522]: DEBUG nova.compute.manager [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 853.781044] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8b4318-06da-4ce5-80d7-ae28e54cd90e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.805341] env[65522]: DEBUG nova.scheduler.client.report [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 853.814765] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.814985] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 853.815266] env[65522]: DEBUG nova.network.neutron [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 853.816610] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114129, 'name': ReconfigVM_Task, 'duration_secs': 1.325396} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.819175] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 7617189c-a902-42e7-8165-0e7c4a1de06d/7617189c-a902-42e7-8165-0e7c4a1de06d.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 853.819175] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7947d346-1a33-493f-ae71-66ff6f40165f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.829072] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 853.829072] env[65522]: value = "task-5114132" [ 853.829072] env[65522]: _type = "Task" [ 853.829072] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.844734] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114132, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.887529] env[65522]: WARNING neutronclient.v2_0.client [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 853.890286] env[65522]: WARNING openstack [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 853.890286] env[65522]: WARNING openstack [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 854.116781] env[65522]: DEBUG oslo_concurrency.lockutils [req-718346e1-bdca-484c-bef3-763ce5af0838 req-ae32fdb2-2618-4b69-a80f-0d852a1a52f4 service nova] Releasing lock "refresh_cache-7617189c-a902-42e7-8165-0e7c4a1de06d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 854.229311] env[65522]: DEBUG nova.compute.manager [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 854.252059] env[65522]: DEBUG nova.network.neutron [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Successfully created port: 30b8dee9-9bde-4601-af20-2015c5e9526d {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 854.259904] env[65522]: DEBUG nova.virt.hardware [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 854.259904] env[65522]: DEBUG nova.virt.hardware [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 854.259904] env[65522]: DEBUG nova.virt.hardware [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 854.260159] env[65522]: DEBUG nova.virt.hardware [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 854.260875] env[65522]: DEBUG nova.virt.hardware [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 854.260875] env[65522]: DEBUG nova.virt.hardware [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 854.260875] env[65522]: DEBUG nova.virt.hardware [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.261128] env[65522]: DEBUG nova.virt.hardware [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 854.261128] env[65522]: DEBUG nova.virt.hardware [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 854.262069] env[65522]: DEBUG nova.virt.hardware [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 854.262069] env[65522]: DEBUG nova.virt.hardware [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 854.262347] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f3af6b-4ae6-4a9b-9094-4b1599f19957 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.273372] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47181507-6512-4a17-a8b2-d0f25d39b1eb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.308625] env[65522]: INFO nova.compute.manager [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Took 46.10 seconds to build instance. [ 854.320608] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.124s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 854.327273] env[65522]: WARNING openstack [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 854.328057] env[65522]: WARNING openstack [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 854.342190] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.280s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 854.343832] env[65522]: INFO nova.compute.claims [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.359409] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114132, 'name': Rename_Task, 'duration_secs': 0.205852} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.359717] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 854.360158] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-28b4d1ff-cb8a-45b7-90aa-ba92ad697f80 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.369399] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 854.369399] env[65522]: value = "task-5114133" [ 854.369399] env[65522]: _type = "Task" [ 854.369399] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.374827] env[65522]: INFO nova.scheduler.client.report [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Deleted allocations for instance 1be995d0-3141-4dcd-ad9a-278d9874e457 [ 854.382106] env[65522]: DEBUG nova.network.neutron [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 854.392896] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114133, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.576134] env[65522]: WARNING openstack [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 854.576563] env[65522]: WARNING openstack [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 854.609937] env[65522]: WARNING openstack [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 854.610121] env[65522]: WARNING openstack [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 854.811129] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cf194f3-0603-48b0-8ccc-beb648a2f8d0 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Lock "26432ad6-610e-49c2-8699-f1e7e128e5b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.619s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 854.882884] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114133, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.883838] env[65522]: DEBUG oslo_concurrency.lockutils [None req-17eb12d4-dfdb-4162-8368-fc2168a9f78c tempest-ServerDiagnosticsV248Test-13148313 tempest-ServerDiagnosticsV248Test-13148313-project-member] Lock "1be995d0-3141-4dcd-ad9a-278d9874e457" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.874s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 854.943836] env[65522]: DEBUG nova.compute.manager [req-03e4e44b-2141-448c-aed3-3ba21474c530 req-793bef3c-e3f5-4042-a6cb-72a293d99a5d service nova] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Received event network-vif-deleted-20608ea3-7885-4006-82d3-66674cbd0e61 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 855.200273] env[65522]: WARNING neutronclient.v2_0.client [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 855.200978] env[65522]: WARNING openstack [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.201344] env[65522]: WARNING openstack [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.293086] env[65522]: WARNING neutronclient.v2_0.client [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 855.293086] env[65522]: WARNING openstack [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.293086] env[65522]: WARNING openstack [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.391745] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114133, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.774159] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e662db0f-d1ca-46bc-a01c-24c1fe413516 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.782808] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494407fb-efb1-4e44-b1b8-1699048a0098 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.823579] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f24a830-c6f9-4aa7-b4ef-2bb52ea1dd77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.833785] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289e84f5-1038-4a30-894f-d48a71db0ef5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.852337] env[65522]: DEBUG nova.compute.provider_tree [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.856616] env[65522]: DEBUG nova.network.neutron [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating instance_info_cache with network_info: [{"id": "9446c1bb-b443-4a52-879d-5fa6cb83f0dc", "address": "fa:16:3e:54:30:7c", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9446c1bb-b4", "ovs_interfaceid": "9446c1bb-b443-4a52-879d-5fa6cb83f0dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 855.860526] env[65522]: DEBUG nova.network.neutron [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Successfully updated port: 30b8dee9-9bde-4601-af20-2015c5e9526d {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 855.884203] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114133, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.941015] env[65522]: DEBUG nova.network.neutron [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance_info_cache with network_info: [{"id": "3f79042c-23dd-4fc0-b3cf-1e2884edabc1", "address": "fa:16:3e:f7:f5:9e", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.131", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f79042c-23", "ovs_interfaceid": "3f79042c-23dd-4fc0-b3cf-1e2884edabc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 856.161209] env[65522]: DEBUG nova.compute.manager [req-2264139b-5a78-4dbe-a970-3dced297ff12 req-b38dfe3c-274a-4f6f-950a-e290ee401da3 service nova] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Received event network-vif-plugged-3f79042c-23dd-4fc0-b3cf-1e2884edabc1 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 856.161483] env[65522]: DEBUG oslo_concurrency.lockutils [req-2264139b-5a78-4dbe-a970-3dced297ff12 req-b38dfe3c-274a-4f6f-950a-e290ee401da3 service nova] Acquiring lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 856.161685] env[65522]: DEBUG oslo_concurrency.lockutils [req-2264139b-5a78-4dbe-a970-3dced297ff12 req-b38dfe3c-274a-4f6f-950a-e290ee401da3 service nova] Lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.161915] env[65522]: DEBUG oslo_concurrency.lockutils [req-2264139b-5a78-4dbe-a970-3dced297ff12 req-b38dfe3c-274a-4f6f-950a-e290ee401da3 service nova] Lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 856.162068] env[65522]: DEBUG nova.compute.manager [req-2264139b-5a78-4dbe-a970-3dced297ff12 req-b38dfe3c-274a-4f6f-950a-e290ee401da3 service nova] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] No waiting events found dispatching network-vif-plugged-3f79042c-23dd-4fc0-b3cf-1e2884edabc1 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 856.162236] env[65522]: WARNING nova.compute.manager [req-2264139b-5a78-4dbe-a970-3dced297ff12 req-b38dfe3c-274a-4f6f-950a-e290ee401da3 service nova] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Received unexpected event network-vif-plugged-3f79042c-23dd-4fc0-b3cf-1e2884edabc1 for instance with vm_state building and task_state spawning. [ 856.360619] env[65522]: DEBUG nova.scheduler.client.report [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 856.366584] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Acquiring lock "refresh_cache-8eda6816-6781-4ad9-9210-11fa623f24b0" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.366584] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Acquired lock "refresh_cache-8eda6816-6781-4ad9-9210-11fa623f24b0" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 856.366584] env[65522]: DEBUG nova.network.neutron [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 856.369722] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 856.386045] env[65522]: DEBUG oslo_vmware.api [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114133, 'name': PowerOnVM_Task, 'duration_secs': 1.767866} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.386045] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.386045] env[65522]: INFO nova.compute.manager [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Took 10.34 seconds to spawn the instance on the hypervisor. [ 856.386045] env[65522]: DEBUG nova.compute.manager [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 856.386604] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4552e072-dec2-4f91-9290-edc7eda6b4c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.447041] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Releasing lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 856.447041] env[65522]: DEBUG nova.compute.manager [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Instance network_info: |[{"id": "3f79042c-23dd-4fc0-b3cf-1e2884edabc1", "address": "fa:16:3e:f7:f5:9e", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.131", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f79042c-23", "ovs_interfaceid": "3f79042c-23dd-4fc0-b3cf-1e2884edabc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 856.447041] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:f5:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3f79042c-23dd-4fc0-b3cf-1e2884edabc1', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.456631] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 856.457789] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 856.458058] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-351c9c79-2f9f-4ad3-863a-04a1b9ff3b00 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.480822] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.480822] env[65522]: value = "task-5114134" [ 856.480822] env[65522]: _type = "Task" [ 856.480822] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.491836] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114134, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.874413] env[65522]: WARNING openstack [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 856.874413] env[65522]: WARNING openstack [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 856.888393] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.545s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 856.888393] env[65522]: DEBUG nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 856.895219] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.809s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.896958] env[65522]: INFO nova.compute.claims [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 856.911076] env[65522]: INFO nova.compute.manager [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Took 53.04 seconds to build instance. [ 856.918255] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7573fd3-d9c0-4059-be9b-3301f1ab5584 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.944020] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ca77354-87ab-4bf4-8292-c9b2d908821a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.955485] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating instance '495fa98f-a9db-4214-87cc-d29209d3cb62' progress to 83 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 856.993085] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114134, 'name': CreateVM_Task, 'duration_secs': 0.50738} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.994517] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 856.994626] env[65522]: WARNING neutronclient.v2_0.client [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 856.995024] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.995311] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 856.995592] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 856.996225] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d351f216-3164-4c98-868c-bed0f83cb553 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.002367] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 857.002367] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522cce65-7350-e3a3-d721-862a25996ae1" [ 857.002367] env[65522]: _type = "Task" [ 857.002367] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.012854] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522cce65-7350-e3a3-d721-862a25996ae1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.163959] env[65522]: DEBUG nova.network.neutron [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 857.403281] env[65522]: DEBUG nova.compute.utils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 857.405910] env[65522]: DEBUG nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 857.405910] env[65522]: DEBUG nova.network.neutron [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 857.405910] env[65522]: WARNING neutronclient.v2_0.client [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 857.406241] env[65522]: WARNING neutronclient.v2_0.client [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 857.410751] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 857.410948] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 857.420814] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ebabea63-15c1-4e00-8824-fd74ec897d22 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "7617189c-a902-42e7-8165-0e7c4a1de06d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.147s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 857.462999] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 857.463344] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e7ad4398-1a00-4ec4-9085-4eb27f7edbdb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.471875] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 857.471875] env[65522]: value = "task-5114135" [ 857.471875] env[65522]: _type = "Task" [ 857.471875] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.482900] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114135, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.512144] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "308330f8-1219-4c02-a129-22b29f0e33b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 857.512396] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "308330f8-1219-4c02-a129-22b29f0e33b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 857.521237] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522cce65-7350-e3a3-d721-862a25996ae1, 'name': SearchDatastore_Task, 'duration_secs': 0.016807} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.521237] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 857.521468] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.522281] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.522281] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 857.522562] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.523331] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a590e27c-0d9a-474b-ade2-c38fb7deab77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.534792] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.534976] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 857.536051] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dcf553c-ebb6-4f7a-812a-4cf5ecf562b2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.542473] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 857.542473] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52baf493-46e6-e22d-1f0b-f2af1cf2606b" [ 857.542473] env[65522]: _type = "Task" [ 857.542473] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.553261] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52baf493-46e6-e22d-1f0b-f2af1cf2606b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.718510] env[65522]: WARNING openstack [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 857.718608] env[65522]: WARNING openstack [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 857.918543] env[65522]: DEBUG nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 857.987752] env[65522]: DEBUG oslo_vmware.api [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114135, 'name': PowerOnVM_Task, 'duration_secs': 0.457014} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.991215] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.991956] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bd3a09a1-6598-40ff-a38b-63f35f9ef371 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating instance '495fa98f-a9db-4214-87cc-d29209d3cb62' progress to 100 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 858.015379] env[65522]: DEBUG nova.compute.manager [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 858.064611] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52baf493-46e6-e22d-1f0b-f2af1cf2606b, 'name': SearchDatastore_Task, 'duration_secs': 0.030743} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.070271] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a25f515-7ce0-42b4-8e92-b63b343470d8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.079050] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 858.079050] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5245a4d5-2b60-d3c5-31ed-b61a6080a8e6" [ 858.079050] env[65522]: _type = "Task" [ 858.079050] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.090830] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5245a4d5-2b60-d3c5-31ed-b61a6080a8e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.096952] env[65522]: DEBUG nova.policy [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a48bd1f8ea6b4c2ab92d4308f986f625', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'acbde84c090e4c0c919728b809c73d02', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 858.538460] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 858.569161] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0fc9db-be50-46a0-9b15-22f19dbcc47d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.577916] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e413b98e-5636-44a0-bf8a-1db3b59005b2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.591810] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5245a4d5-2b60-d3c5-31ed-b61a6080a8e6, 'name': SearchDatastore_Task, 'duration_secs': 0.042419} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.625212] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 858.625554] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 5e227f6c-9f13-416f-8e6d-2f7d931619fd/5e227f6c-9f13-416f-8e6d-2f7d931619fd.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 858.626034] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bcc6e2b7-ea07-49ad-8dc8-4205af8b7684 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.629249] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60326a10-75d2-4120-b16c-4884e6ba10a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.637487] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262d125f-b997-46b8-afc2-4208bf4650e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.645666] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 858.645666] env[65522]: value = "task-5114136" [ 858.645666] env[65522]: _type = "Task" [ 858.645666] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.660522] env[65522]: DEBUG nova.compute.provider_tree [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.672400] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114136, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.682261] env[65522]: DEBUG nova.network.neutron [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Successfully created port: dd99a89e-fb55-4af2-94d6-979e1b1cae45 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 858.709088] env[65522]: WARNING neutronclient.v2_0.client [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 858.709904] env[65522]: WARNING openstack [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 858.710271] env[65522]: WARNING openstack [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 858.934893] env[65522]: DEBUG nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 858.971100] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 858.971394] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 858.971533] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 858.971715] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 858.971856] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 858.971997] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 858.972217] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.972917] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 858.972917] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 858.972917] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 858.972917] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 858.973945] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae298c57-c1bf-4519-bc49-836db563ad21 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.984392] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a902a4a1-7016-4e8b-9f8e-7db0d758abbb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.165792] env[65522]: DEBUG nova.scheduler.client.report [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 859.171397] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114136, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.664161] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114136, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.68361} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.664582] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 5e227f6c-9f13-416f-8e6d-2f7d931619fd/5e227f6c-9f13-416f-8e6d-2f7d931619fd.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 859.664697] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.665043] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-70f1782d-68ab-4cd3-b702-18362a0b4c19 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.674070] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.779s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 859.674599] env[65522]: DEBUG nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 859.680837] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 859.680837] env[65522]: value = "task-5114137" [ 859.680837] env[65522]: _type = "Task" [ 859.680837] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.681139] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.570s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 859.681367] env[65522]: DEBUG nova.objects.instance [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lazy-loading 'resources' on Instance uuid da211708-a6a2-4e03-bb41-c5b93564de20 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.697777] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114137, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.191503] env[65522]: DEBUG nova.compute.utils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 860.199565] env[65522]: DEBUG nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 860.199565] env[65522]: DEBUG nova.network.neutron [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 860.199565] env[65522]: WARNING neutronclient.v2_0.client [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 860.199565] env[65522]: WARNING neutronclient.v2_0.client [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 860.199565] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 860.199847] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.216045] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114137, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.17109} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.216157] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.217472] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0090f963-1dc8-4669-9026-8b484a5f96de {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.251844] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 5e227f6c-9f13-416f-8e6d-2f7d931619fd/5e227f6c-9f13-416f-8e6d-2f7d931619fd.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.254974] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3f51c09-c6bc-46ad-931a-237ad5e19d02 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.289577] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 860.289577] env[65522]: value = "task-5114138" [ 860.289577] env[65522]: _type = "Task" [ 860.289577] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.300595] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114138, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.651407] env[65522]: DEBUG nova.network.neutron [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Successfully updated port: dd99a89e-fb55-4af2-94d6-979e1b1cae45 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 860.701254] env[65522]: DEBUG nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 860.710819] env[65522]: DEBUG nova.network.neutron [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Updating instance_info_cache with network_info: [{"id": "30b8dee9-9bde-4601-af20-2015c5e9526d", "address": "fa:16:3e:6e:0a:77", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.161", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30b8dee9-9b", "ovs_interfaceid": "30b8dee9-9bde-4601-af20-2015c5e9526d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 860.783433] env[65522]: DEBUG nova.policy [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a48bd1f8ea6b4c2ab92d4308f986f625', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'acbde84c090e4c0c919728b809c73d02', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 860.818329] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114138, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.820765] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc5d0d7f-43c4-4dcc-a076-8d31c909414d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.829358] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ab2c94-3364-4c0f-a4ec-4a2fa693735f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.862450] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92732005-79f6-410c-93cf-174cc8b76825 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.878206] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f00aa291-5d42-467c-8120-f4111ce071d6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.899054] env[65522]: DEBUG nova.compute.provider_tree [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.065818] env[65522]: DEBUG nova.compute.manager [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Received event network-changed-cb206964-4331-495a-94ce-c2b95dccc600 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 861.065818] env[65522]: DEBUG nova.compute.manager [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Refreshing instance network info cache due to event network-changed-cb206964-4331-495a-94ce-c2b95dccc600. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 861.065818] env[65522]: DEBUG oslo_concurrency.lockutils [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] Acquiring lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.065818] env[65522]: DEBUG oslo_concurrency.lockutils [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] Acquired lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 861.065818] env[65522]: DEBUG nova.network.neutron [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Refreshing network info cache for port cb206964-4331-495a-94ce-c2b95dccc600 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 861.153546] env[65522]: DEBUG nova.compute.manager [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Received event network-changed-3f79042c-23dd-4fc0-b3cf-1e2884edabc1 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 861.153546] env[65522]: DEBUG nova.compute.manager [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Refreshing instance network info cache due to event network-changed-3f79042c-23dd-4fc0-b3cf-1e2884edabc1. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 861.153546] env[65522]: DEBUG oslo_concurrency.lockutils [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Acquiring lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.153546] env[65522]: DEBUG oslo_concurrency.lockutils [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Acquired lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 861.153546] env[65522]: DEBUG nova.network.neutron [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Refreshing network info cache for port 3f79042c-23dd-4fc0-b3cf-1e2884edabc1 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 861.156960] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "refresh_cache-179a37ed-1176-4082-8c68-1be6abd3fd9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.158730] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired lock "refresh_cache-179a37ed-1176-4082-8c68-1be6abd3fd9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 861.158730] env[65522]: DEBUG nova.network.neutron [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 861.189989] env[65522]: DEBUG nova.network.neutron [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Successfully created port: e296c3b2-a6e6-4f3a-8fc8-9a208bb85470 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 861.218527] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Releasing lock "refresh_cache-8eda6816-6781-4ad9-9210-11fa623f24b0" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 861.218527] env[65522]: DEBUG nova.compute.manager [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Instance network_info: |[{"id": "30b8dee9-9bde-4601-af20-2015c5e9526d", "address": "fa:16:3e:6e:0a:77", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.161", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30b8dee9-9b", "ovs_interfaceid": "30b8dee9-9bde-4601-af20-2015c5e9526d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 861.219810] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:0a:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc0e97b-b21d-4557-a4d4-fd7e8f973368', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '30b8dee9-9bde-4601-af20-2015c5e9526d', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.227861] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Creating folder: Project (b5c2d36708f44bb9a0aa34f49014a3ed). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.228800] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2f9cdfeb-ebe7-4d0f-b96e-05bb7b2af288 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.244383] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Created folder: Project (b5c2d36708f44bb9a0aa34f49014a3ed) in parent group-v994660. [ 861.244563] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Creating folder: Instances. Parent ref: group-v994818. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.244900] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-59c384a7-6c3c-4b41-ae7c-cfbd1acf807b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.259063] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Created folder: Instances in parent group-v994818. [ 861.259354] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 861.259573] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 861.259862] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1a93ab7c-9e0f-48e6-ad3c-c605395f6e80 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.281913] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.281913] env[65522]: value = "task-5114141" [ 861.281913] env[65522]: _type = "Task" [ 861.281913] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.294274] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114141, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.305473] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114138, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.402887] env[65522]: DEBUG nova.scheduler.client.report [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 861.524935] env[65522]: INFO nova.compute.manager [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Rebuilding instance [ 861.569404] env[65522]: WARNING neutronclient.v2_0.client [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.570132] env[65522]: WARNING openstack [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.570217] env[65522]: WARNING openstack [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.598840] env[65522]: DEBUG nova.compute.manager [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 861.598840] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a78f90-f5b6-4495-ac9b-5f29a15e2921 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.659458] env[65522]: WARNING neutronclient.v2_0.client [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.660365] env[65522]: WARNING openstack [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.660726] env[65522]: WARNING openstack [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.672539] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.674338] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.731057] env[65522]: DEBUG nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 861.757169] env[65522]: DEBUG nova.network.neutron [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 861.767264] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 861.767264] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 861.767264] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 861.767422] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 861.767491] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 861.767766] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 861.767992] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.768165] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 861.768327] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 861.768486] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 861.769048] env[65522]: DEBUG nova.virt.hardware [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 861.769999] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd81b2b-ac44-4aae-a6c2-38896735a733 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.774103] env[65522]: WARNING openstack [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.774470] env[65522]: WARNING openstack [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.793728] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d74b99-6073-4ef5-9fe6-54f9f421f971 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.803920] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114141, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.824620] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114138, 'name': ReconfigVM_Task, 'duration_secs': 1.024649} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.824913] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 5e227f6c-9f13-416f-8e6d-2f7d931619fd/5e227f6c-9f13-416f-8e6d-2f7d931619fd.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.825609] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c8871710-fd05-43d2-852a-a85a089705f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.837289] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 861.837289] env[65522]: value = "task-5114142" [ 861.837289] env[65522]: _type = "Task" [ 861.837289] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.849346] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114142, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.894315] env[65522]: DEBUG oslo_concurrency.lockutils [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 861.894722] env[65522]: DEBUG oslo_concurrency.lockutils [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.002s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 861.894852] env[65522]: INFO nova.compute.manager [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Rebooting instance [ 861.920190] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.235s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 861.922353] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.515s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 861.924587] env[65522]: INFO nova.compute.claims [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.932834] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.935797] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.951997] env[65522]: INFO nova.scheduler.client.report [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Deleted allocations for instance da211708-a6a2-4e03-bb41-c5b93564de20 [ 861.993156] env[65522]: WARNING openstack [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.993156] env[65522]: WARNING openstack [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.296055] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114141, 'name': CreateVM_Task, 'duration_secs': 0.553113} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.296365] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 862.298377] env[65522]: WARNING neutronclient.v2_0.client [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 862.298897] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.299073] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 862.299465] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 862.299801] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d62c9b85-cdd4-4694-9616-cce8a7aa5f3d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.306404] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Waiting for the task: (returnval){ [ 862.306404] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52060997-90c0-de35-752b-4a2a1da46927" [ 862.306404] env[65522]: _type = "Task" [ 862.306404] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.316235] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52060997-90c0-de35-752b-4a2a1da46927, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.347842] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114142, 'name': Rename_Task, 'duration_secs': 0.240399} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.348129] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.348428] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13c0423e-be32-4285-8a50-82bdf28eda23 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.358549] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "4264998b-949d-4b01-8832-fb1df6214403" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.358783] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "4264998b-949d-4b01-8832-fb1df6214403" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 862.359974] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 862.359974] env[65522]: value = "task-5114143" [ 862.359974] env[65522]: _type = "Task" [ 862.359974] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.371011] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114143, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.423280] env[65522]: DEBUG oslo_concurrency.lockutils [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.462470] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10858f93-0099-400a-904d-d4bc112d1a41 tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "da211708-a6a2-4e03-bb41-c5b93564de20" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.900s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 862.573772] env[65522]: WARNING neutronclient.v2_0.client [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 862.574659] env[65522]: WARNING openstack [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.574808] env[65522]: WARNING openstack [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.614599] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 862.615748] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ca0bc59-cd95-41dc-9cfa-762e4bd0879c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.627946] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 862.627946] env[65522]: value = "task-5114144" [ 862.627946] env[65522]: _type = "Task" [ 862.627946] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.639411] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114144, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.677663] env[65522]: WARNING neutronclient.v2_0.client [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 862.678163] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.678873] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.746124] env[65522]: WARNING neutronclient.v2_0.client [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 862.749414] env[65522]: WARNING openstack [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.749414] env[65522]: WARNING openstack [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.817844] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52060997-90c0-de35-752b-4a2a1da46927, 'name': SearchDatastore_Task, 'duration_secs': 0.014807} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.823357] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 862.823637] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 862.823907] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.824111] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 862.824355] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 862.824959] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd4a2ee0-ca23-4a53-acca-b5d9707c4e02 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.835906] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 862.836390] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 862.837042] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60c9eacd-3dd3-4034-840e-cfaa11e02abb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.846247] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Waiting for the task: (returnval){ [ 862.846247] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528d63ff-fbba-285a-13ca-cdca174ab14b" [ 862.846247] env[65522]: _type = "Task" [ 862.846247] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.862473] env[65522]: DEBUG nova.compute.manager [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 862.864860] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528d63ff-fbba-285a-13ca-cdca174ab14b, 'name': SearchDatastore_Task, 'duration_secs': 0.011915} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.870032] env[65522]: DEBUG nova.network.neutron [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Successfully updated port: e296c3b2-a6e6-4f3a-8fc8-9a208bb85470 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 862.872024] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7acfdaef-838e-4de4-ac88-99f2aa702fe4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.881781] env[65522]: DEBUG nova.network.neutron [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Updated VIF entry in instance network info cache for port cb206964-4331-495a-94ce-c2b95dccc600. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 862.882182] env[65522]: DEBUG nova.network.neutron [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Updating instance_info_cache with network_info: [{"id": "cb206964-4331-495a-94ce-c2b95dccc600", "address": "fa:16:3e:5b:a0:80", "network": {"id": "922eeeeb-e124-4ae5-97a4-0e995b16f7ec", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-437359683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37fc01876567476f9b93d765b2cfddc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccf76700-491b-4462-ab19-e6d3a9ff87ac", "external-id": "nsx-vlan-transportzone-956", "segmentation_id": 956, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb206964-43", "ovs_interfaceid": "cb206964-4331-495a-94ce-c2b95dccc600", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 862.889869] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Waiting for the task: (returnval){ [ 862.889869] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52561cd9-db3b-7415-f1a6-3e147bd0a6bd" [ 862.889869] env[65522]: _type = "Task" [ 862.889869] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.890197] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114143, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.903410] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52561cd9-db3b-7415-f1a6-3e147bd0a6bd, 'name': SearchDatastore_Task, 'duration_secs': 0.0124} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.903410] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 862.903638] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 8eda6816-6781-4ad9-9210-11fa623f24b0/8eda6816-6781-4ad9-9210-11fa623f24b0.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 862.903954] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-924cd10e-0ba2-43fa-8400-bed24faab2e3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.912114] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Waiting for the task: (returnval){ [ 862.912114] env[65522]: value = "task-5114145" [ 862.912114] env[65522]: _type = "Task" [ 862.912114] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.924454] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114145, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.962612] env[65522]: DEBUG nova.network.neutron [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Updating instance_info_cache with network_info: [{"id": "dd99a89e-fb55-4af2-94d6-979e1b1cae45", "address": "fa:16:3e:21:34:eb", "network": {"id": "0ab23355-59a7-409b-89f3-089fbd03db5e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-960579569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acbde84c090e4c0c919728b809c73d02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd99a89e-fb", "ovs_interfaceid": "dd99a89e-fb55-4af2-94d6-979e1b1cae45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 863.009468] env[65522]: DEBUG nova.network.neutron [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updated VIF entry in instance network info cache for port 3f79042c-23dd-4fc0-b3cf-1e2884edabc1. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 863.009957] env[65522]: DEBUG nova.network.neutron [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance_info_cache with network_info: [{"id": "3f79042c-23dd-4fc0-b3cf-1e2884edabc1", "address": "fa:16:3e:f7:f5:9e", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.131", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f79042c-23", "ovs_interfaceid": "3f79042c-23dd-4fc0-b3cf-1e2884edabc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 863.147177] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114144, 'name': PowerOffVM_Task, 'duration_secs': 0.133031} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.150199] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 863.150700] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 863.152190] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e96d28-0067-4c01-beb6-996d2f54e7a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.163422] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 863.163897] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-492f8f09-7e55-473d-ad0b-f375108b4bfe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.197046] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 863.197360] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 863.197520] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Deleting the datastore file [datastore1] 26432ad6-610e-49c2-8699-f1e7e128e5b8 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.197865] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-56f93b36-dd5f-4c6d-aabf-fefaac11e256 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.217770] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 863.217770] env[65522]: value = "task-5114147" [ 863.217770] env[65522]: _type = "Task" [ 863.217770] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.234305] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114147, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.356287] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "495fa98f-a9db-4214-87cc-d29209d3cb62" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.356571] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "495fa98f-a9db-4214-87cc-d29209d3cb62" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.356763] env[65522]: DEBUG nova.compute.manager [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Going to confirm migration 2 {{(pid=65522) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 863.376979] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "refresh_cache-6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.377882] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired lock "refresh_cache-6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.377882] env[65522]: DEBUG nova.network.neutron [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 863.382669] env[65522]: DEBUG oslo_vmware.api [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114143, 'name': PowerOnVM_Task, 'duration_secs': 0.761618} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.383451] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 863.383567] env[65522]: INFO nova.compute.manager [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Took 11.88 seconds to spawn the instance on the hypervisor. [ 863.383689] env[65522]: DEBUG nova.compute.manager [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 863.384841] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6df956-7186-4855-8a66-21869f4b3f71 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.395432] env[65522]: DEBUG oslo_concurrency.lockutils [req-375fdadc-027b-4b96-9c32-1089cb81cf09 req-adb3bb28-9da8-442f-b393-bac1ff917876 service nova] Releasing lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 863.399480] env[65522]: DEBUG oslo_concurrency.lockutils [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquired lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.399480] env[65522]: DEBUG nova.network.neutron [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 863.404513] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.436268] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114145, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.469085] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Releasing lock "refresh_cache-179a37ed-1176-4082-8c68-1be6abd3fd9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 863.469505] env[65522]: DEBUG nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Instance network_info: |[{"id": "dd99a89e-fb55-4af2-94d6-979e1b1cae45", "address": "fa:16:3e:21:34:eb", "network": {"id": "0ab23355-59a7-409b-89f3-089fbd03db5e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-960579569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acbde84c090e4c0c919728b809c73d02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd99a89e-fb", "ovs_interfaceid": "dd99a89e-fb55-4af2-94d6-979e1b1cae45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 863.470427] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:34:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4712af2-45ef-4652-8d2c-482ec70056d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd99a89e-fb55-4af2-94d6-979e1b1cae45', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 863.479341] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 863.482967] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 863.484841] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8e815563-8fc6-4160-a01b-8218ea084b68 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.504556] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1912b363-648f-4736-af18-e17919ac2ae4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.514696] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f89a35a9-2c95-4802-8d37-0975c9d5f6a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.518430] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 863.518430] env[65522]: value = "task-5114148" [ 863.518430] env[65522]: _type = "Task" [ 863.518430] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.519062] env[65522]: DEBUG oslo_concurrency.lockutils [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Releasing lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 863.519321] env[65522]: DEBUG nova.compute.manager [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Received event network-vif-plugged-30b8dee9-9bde-4601-af20-2015c5e9526d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 863.519518] env[65522]: DEBUG oslo_concurrency.lockutils [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Acquiring lock "8eda6816-6781-4ad9-9210-11fa623f24b0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.519745] env[65522]: DEBUG oslo_concurrency.lockutils [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Lock "8eda6816-6781-4ad9-9210-11fa623f24b0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.519923] env[65522]: DEBUG oslo_concurrency.lockutils [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Lock "8eda6816-6781-4ad9-9210-11fa623f24b0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.520170] env[65522]: DEBUG nova.compute.manager [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] No waiting events found dispatching network-vif-plugged-30b8dee9-9bde-4601-af20-2015c5e9526d {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 863.520346] env[65522]: WARNING nova.compute.manager [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Received unexpected event network-vif-plugged-30b8dee9-9bde-4601-af20-2015c5e9526d for instance with vm_state building and task_state spawning. [ 863.520532] env[65522]: DEBUG nova.compute.manager [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Received event network-changed-30b8dee9-9bde-4601-af20-2015c5e9526d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 863.520707] env[65522]: DEBUG nova.compute.manager [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Refreshing instance network info cache due to event network-changed-30b8dee9-9bde-4601-af20-2015c5e9526d. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 863.520977] env[65522]: DEBUG oslo_concurrency.lockutils [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Acquiring lock "refresh_cache-8eda6816-6781-4ad9-9210-11fa623f24b0" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.521066] env[65522]: DEBUG oslo_concurrency.lockutils [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Acquired lock "refresh_cache-8eda6816-6781-4ad9-9210-11fa623f24b0" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.521199] env[65522]: DEBUG nova.network.neutron [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Refreshing network info cache for port 30b8dee9-9bde-4601-af20-2015c5e9526d {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 863.557629] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf0796b-50f9-49a5-9f17-4172af9c9ed6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.564431] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114148, 'name': CreateVM_Task} progress is 15%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.570951] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d8279f-1013-49b7-8322-fcb71a96241f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.589538] env[65522]: DEBUG nova.compute.provider_tree [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.712985] env[65522]: DEBUG nova.compute.manager [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Received event network-changed-dc729a30-fef3-4b2e-ab41-6c7a4eb89f73 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 863.714721] env[65522]: DEBUG nova.compute.manager [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Refreshing instance network info cache due to event network-changed-dc729a30-fef3-4b2e-ab41-6c7a4eb89f73. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 863.714721] env[65522]: DEBUG oslo_concurrency.lockutils [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] Acquiring lock "refresh_cache-7617189c-a902-42e7-8165-0e7c4a1de06d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.714721] env[65522]: DEBUG oslo_concurrency.lockutils [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] Acquired lock "refresh_cache-7617189c-a902-42e7-8165-0e7c4a1de06d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.714721] env[65522]: DEBUG nova.network.neutron [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Refreshing network info cache for port dc729a30-fef3-4b2e-ab41-6c7a4eb89f73 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 863.721661] env[65522]: DEBUG nova.compute.manager [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Received event network-vif-plugged-dd99a89e-fb55-4af2-94d6-979e1b1cae45 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 863.721925] env[65522]: DEBUG oslo_concurrency.lockutils [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Acquiring lock "179a37ed-1176-4082-8c68-1be6abd3fd9d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.723140] env[65522]: DEBUG oslo_concurrency.lockutils [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Lock "179a37ed-1176-4082-8c68-1be6abd3fd9d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.723140] env[65522]: DEBUG oslo_concurrency.lockutils [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Lock "179a37ed-1176-4082-8c68-1be6abd3fd9d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.723140] env[65522]: DEBUG nova.compute.manager [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] No waiting events found dispatching network-vif-plugged-dd99a89e-fb55-4af2-94d6-979e1b1cae45 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 863.723140] env[65522]: WARNING nova.compute.manager [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Received unexpected event network-vif-plugged-dd99a89e-fb55-4af2-94d6-979e1b1cae45 for instance with vm_state building and task_state spawning. [ 863.723140] env[65522]: DEBUG nova.compute.manager [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Received event network-changed-dd99a89e-fb55-4af2-94d6-979e1b1cae45 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 863.723537] env[65522]: DEBUG nova.compute.manager [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Refreshing instance network info cache due to event network-changed-dd99a89e-fb55-4af2-94d6-979e1b1cae45. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 863.723794] env[65522]: DEBUG oslo_concurrency.lockutils [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Acquiring lock "refresh_cache-179a37ed-1176-4082-8c68-1be6abd3fd9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.723980] env[65522]: DEBUG oslo_concurrency.lockutils [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Acquired lock "refresh_cache-179a37ed-1176-4082-8c68-1be6abd3fd9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.724217] env[65522]: DEBUG nova.network.neutron [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Refreshing network info cache for port dd99a89e-fb55-4af2-94d6-979e1b1cae45 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 863.737053] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114147, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.363736} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.738211] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 863.738537] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 863.738722] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 863.862926] env[65522]: WARNING neutronclient.v2_0.client [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 863.874126] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Acquiring lock "9da312a0-c6eb-4903-a7d8-e178aa54d2ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.874126] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Lock "9da312a0-c6eb-4903-a7d8-e178aa54d2ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.884402] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.885588] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.909787] env[65522]: WARNING neutronclient.v2_0.client [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 863.910558] env[65522]: WARNING openstack [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.910983] env[65522]: WARNING openstack [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.922862] env[65522]: INFO nova.compute.manager [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Took 53.90 seconds to build instance. [ 863.934594] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114145, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.551523} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.934878] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 8eda6816-6781-4ad9-9210-11fa623f24b0/8eda6816-6781-4ad9-9210-11fa623f24b0.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 863.935223] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.935511] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd6e6aa6-5b41-4470-9e5e-7dd0409113cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.946270] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Waiting for the task: (returnval){ [ 863.946270] env[65522]: value = "task-5114149" [ 863.946270] env[65522]: _type = "Task" [ 863.946270] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.957035] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114149, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.960334] env[65522]: WARNING neutronclient.v2_0.client [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 863.960736] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.960910] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.961091] env[65522]: DEBUG nova.network.neutron [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 863.961275] env[65522]: DEBUG nova.objects.instance [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lazy-loading 'info_cache' on Instance uuid 495fa98f-a9db-4214-87cc-d29209d3cb62 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.974791] env[65522]: DEBUG nova.network.neutron [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 864.011267] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.012041] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.028164] env[65522]: WARNING neutronclient.v2_0.client [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.031090] env[65522]: WARNING openstack [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.031090] env[65522]: WARNING openstack [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.044076] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114148, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.092866] env[65522]: DEBUG nova.scheduler.client.report [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 864.124751] env[65522]: WARNING openstack [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.125165] env[65522]: WARNING openstack [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.134849] env[65522]: WARNING neutronclient.v2_0.client [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.135803] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.136302] env[65522]: WARNING openstack [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.218043] env[65522]: WARNING neutronclient.v2_0.client [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.218254] env[65522]: WARNING openstack [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.219961] env[65522]: WARNING openstack [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.230521] env[65522]: WARNING openstack [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.230939] env[65522]: WARNING openstack [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.243393] env[65522]: WARNING neutronclient.v2_0.client [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.244295] env[65522]: WARNING openstack [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.244923] env[65522]: WARNING openstack [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.262548] env[65522]: WARNING neutronclient.v2_0.client [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.263478] env[65522]: WARNING openstack [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.263987] env[65522]: WARNING openstack [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.296245] env[65522]: DEBUG nova.network.neutron [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Updating instance_info_cache with network_info: [{"id": "e296c3b2-a6e6-4f3a-8fc8-9a208bb85470", "address": "fa:16:3e:31:c2:f0", "network": {"id": "0ab23355-59a7-409b-89f3-089fbd03db5e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-960579569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acbde84c090e4c0c919728b809c73d02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape296c3b2-a6", "ovs_interfaceid": "e296c3b2-a6e6-4f3a-8fc8-9a208bb85470", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 864.374603] env[65522]: DEBUG oslo_concurrency.lockutils [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "361e54c7-4309-48b3-ad3e-b2ebc783741c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 864.374993] env[65522]: DEBUG oslo_concurrency.lockutils [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "361e54c7-4309-48b3-ad3e-b2ebc783741c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 864.375092] env[65522]: DEBUG oslo_concurrency.lockutils [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "361e54c7-4309-48b3-ad3e-b2ebc783741c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 864.375280] env[65522]: DEBUG oslo_concurrency.lockutils [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "361e54c7-4309-48b3-ad3e-b2ebc783741c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 864.375702] env[65522]: DEBUG oslo_concurrency.lockutils [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "361e54c7-4309-48b3-ad3e-b2ebc783741c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 864.386336] env[65522]: DEBUG nova.compute.manager [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 864.390074] env[65522]: INFO nova.compute.manager [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Terminating instance [ 864.408729] env[65522]: WARNING neutronclient.v2_0.client [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.409312] env[65522]: WARNING openstack [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.409656] env[65522]: WARNING openstack [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.428016] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9d88aaef-3d35-456f-9d97-f8aaaff15001 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.419s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 864.445894] env[65522]: DEBUG nova.network.neutron [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Updating instance_info_cache with network_info: [{"id": "cb206964-4331-495a-94ce-c2b95dccc600", "address": "fa:16:3e:5b:a0:80", "network": {"id": "922eeeeb-e124-4ae5-97a4-0e995b16f7ec", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-437359683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37fc01876567476f9b93d765b2cfddc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccf76700-491b-4462-ab19-e6d3a9ff87ac", "external-id": "nsx-vlan-transportzone-956", "segmentation_id": 956, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb206964-43", "ovs_interfaceid": "cb206964-4331-495a-94ce-c2b95dccc600", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 864.457455] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114149, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.126436} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.457762] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.462898] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d62b224-dc64-42ce-90db-91c7a86b069d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.495046] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 8eda6816-6781-4ad9-9210-11fa623f24b0/8eda6816-6781-4ad9-9210-11fa623f24b0.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.498963] env[65522]: WARNING openstack [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.498963] env[65522]: WARNING openstack [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.506470] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1e6189c-277c-430f-ab0d-7c9ae3238a16 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.525175] env[65522]: WARNING openstack [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.525623] env[65522]: WARNING openstack [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.549165] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114148, 'name': CreateVM_Task, 'duration_secs': 0.586904} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.550536] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 864.551406] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Waiting for the task: (returnval){ [ 864.551406] env[65522]: value = "task-5114150" [ 864.551406] env[65522]: _type = "Task" [ 864.551406] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.554467] env[65522]: WARNING neutronclient.v2_0.client [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.554467] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.554467] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 864.554467] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 864.554467] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd7f4e05-cd94-4cbf-bbf1-20c609bb94ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.565173] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 864.565173] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52868e1c-09db-1a0a-2c21-b1301ca7e9eb" [ 864.565173] env[65522]: _type = "Task" [ 864.565173] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.565973] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114150, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.581270] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52868e1c-09db-1a0a-2c21-b1301ca7e9eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.601890] env[65522]: DEBUG nova.network.neutron [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Updated VIF entry in instance network info cache for port 30b8dee9-9bde-4601-af20-2015c5e9526d. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 864.602510] env[65522]: DEBUG nova.network.neutron [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Updating instance_info_cache with network_info: [{"id": "30b8dee9-9bde-4601-af20-2015c5e9526d", "address": "fa:16:3e:6e:0a:77", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.161", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30b8dee9-9b", "ovs_interfaceid": "30b8dee9-9bde-4601-af20-2015c5e9526d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 864.604449] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.682s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 864.604722] env[65522]: DEBUG nova.compute.manager [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 864.608292] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.978s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 864.610034] env[65522]: INFO nova.compute.claims [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.630263] env[65522]: WARNING neutronclient.v2_0.client [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.630806] env[65522]: WARNING openstack [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.631288] env[65522]: WARNING openstack [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.669858] env[65522]: WARNING neutronclient.v2_0.client [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.670594] env[65522]: WARNING openstack [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.671013] env[65522]: WARNING openstack [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.778142] env[65522]: DEBUG nova.network.neutron [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Updated VIF entry in instance network info cache for port dc729a30-fef3-4b2e-ab41-6c7a4eb89f73. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 864.778142] env[65522]: DEBUG nova.network.neutron [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Updating instance_info_cache with network_info: [{"id": "dc729a30-fef3-4b2e-ab41-6c7a4eb89f73", "address": "fa:16:3e:6a:3b:a6", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc729a30-fe", "ovs_interfaceid": "dc729a30-fef3-4b2e-ab41-6c7a4eb89f73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 864.801832] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Releasing lock "refresh_cache-6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 864.802717] env[65522]: DEBUG nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Instance network_info: |[{"id": "e296c3b2-a6e6-4f3a-8fc8-9a208bb85470", "address": "fa:16:3e:31:c2:f0", "network": {"id": "0ab23355-59a7-409b-89f3-089fbd03db5e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-960579569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acbde84c090e4c0c919728b809c73d02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape296c3b2-a6", "ovs_interfaceid": "e296c3b2-a6e6-4f3a-8fc8-9a208bb85470", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 864.803290] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:c2:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4712af2-45ef-4652-8d2c-482ec70056d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e296c3b2-a6e6-4f3a-8fc8-9a208bb85470', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 864.812987] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 864.815645] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 864.816732] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d98baf6f-1075-4a40-9a1c-64106c82c017 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.835921] env[65522]: DEBUG nova.virt.hardware [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 864.836451] env[65522]: DEBUG nova.virt.hardware [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 864.836451] env[65522]: DEBUG nova.virt.hardware [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 864.836581] env[65522]: DEBUG nova.virt.hardware [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 864.836627] env[65522]: DEBUG nova.virt.hardware [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 864.839021] env[65522]: DEBUG nova.virt.hardware [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 864.839021] env[65522]: DEBUG nova.virt.hardware [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 864.839021] env[65522]: DEBUG nova.virt.hardware [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 864.839021] env[65522]: DEBUG nova.virt.hardware [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 864.839021] env[65522]: DEBUG nova.virt.hardware [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 864.839021] env[65522]: DEBUG nova.virt.hardware [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 864.840110] env[65522]: DEBUG nova.network.neutron [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Updated VIF entry in instance network info cache for port dd99a89e-fb55-4af2-94d6-979e1b1cae45. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 864.840456] env[65522]: DEBUG nova.network.neutron [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Updating instance_info_cache with network_info: [{"id": "dd99a89e-fb55-4af2-94d6-979e1b1cae45", "address": "fa:16:3e:21:34:eb", "network": {"id": "0ab23355-59a7-409b-89f3-089fbd03db5e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-960579569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acbde84c090e4c0c919728b809c73d02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd99a89e-fb", "ovs_interfaceid": "dd99a89e-fb55-4af2-94d6-979e1b1cae45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 864.842426] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d22eab-4231-4512-9653-b396da390e3c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.854140] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-724c36c9-277e-480f-9aee-19d49e82e024 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.863508] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 864.863508] env[65522]: value = "task-5114151" [ 864.863508] env[65522]: _type = "Task" [ 864.863508] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.875784] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 864.881875] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 864.883096] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 864.883453] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aeb3b5a4-bcb4-4177-9b13-b08eefdc3b56 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.905572] env[65522]: DEBUG nova.compute.manager [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 864.905572] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 864.905694] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114151, 'name': CreateVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.909516] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2803fc6-1f15-4f90-bfce-7b8fe095c74a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.918496] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.920408] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db29d580-07a8-414f-bff5-56441fe7f001 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.921845] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 864.921845] env[65522]: value = "task-5114152" [ 864.921845] env[65522]: _type = "Task" [ 864.921845] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.930160] env[65522]: DEBUG oslo_vmware.api [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 864.930160] env[65522]: value = "task-5114153" [ 864.930160] env[65522]: _type = "Task" [ 864.930160] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.937144] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114152, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.938508] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 864.946505] env[65522]: DEBUG oslo_vmware.api [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114153, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.952853] env[65522]: DEBUG oslo_concurrency.lockutils [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Releasing lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 864.971013] env[65522]: WARNING neutronclient.v2_0.client [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.971849] env[65522]: WARNING openstack [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.972318] env[65522]: WARNING openstack [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.066448] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114150, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.079668] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52868e1c-09db-1a0a-2c21-b1301ca7e9eb, 'name': SearchDatastore_Task, 'duration_secs': 0.022945} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.080174] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.080335] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.080565] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.080731] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.080971] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.081293] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13b0412b-384f-42f1-9990-8be3d87f6974 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.092930] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.093314] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 865.097027] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a78ef7b4-7468-4446-aed8-1fb03f0921ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.102917] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 865.102917] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52afb45b-cf37-9943-1f04-581a02483d62" [ 865.102917] env[65522]: _type = "Task" [ 865.102917] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.108309] env[65522]: DEBUG oslo_concurrency.lockutils [req-99c4afa1-ad39-4ffa-9b41-7f712b0af092 req-8483f099-e994-4245-8de8-078a8c45b2e7 service nova] Releasing lock "refresh_cache-8eda6816-6781-4ad9-9210-11fa623f24b0" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.115328] env[65522]: DEBUG nova.compute.utils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 865.119353] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52afb45b-cf37-9943-1f04-581a02483d62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.120136] env[65522]: DEBUG nova.compute.manager [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 865.120340] env[65522]: DEBUG nova.network.neutron [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 865.120701] env[65522]: WARNING neutronclient.v2_0.client [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.121058] env[65522]: WARNING neutronclient.v2_0.client [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.121757] env[65522]: WARNING openstack [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.122124] env[65522]: WARNING openstack [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.181968] env[65522]: DEBUG nova.policy [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b491939798e4481fb433ffb81f366b25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '068c2387de8c406194d9b1762c7292a7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 865.190065] env[65522]: WARNING openstack [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.190591] env[65522]: WARNING openstack [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.282469] env[65522]: DEBUG oslo_concurrency.lockutils [req-43c2da26-4ba8-4663-baef-a64d2d7edf51 req-dba6f9f9-587a-4536-800f-48ea63bff883 service nova] Releasing lock "refresh_cache-7617189c-a902-42e7-8165-0e7c4a1de06d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.286401] env[65522]: WARNING neutronclient.v2_0.client [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.287128] env[65522]: WARNING openstack [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.287556] env[65522]: WARNING openstack [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.347376] env[65522]: DEBUG oslo_concurrency.lockutils [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Releasing lock "refresh_cache-179a37ed-1176-4082-8c68-1be6abd3fd9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.347929] env[65522]: DEBUG nova.compute.manager [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Received event network-vif-plugged-e296c3b2-a6e6-4f3a-8fc8-9a208bb85470 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 865.347929] env[65522]: DEBUG oslo_concurrency.lockutils [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Acquiring lock "6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 865.348115] env[65522]: DEBUG oslo_concurrency.lockutils [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Lock "6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 865.348256] env[65522]: DEBUG oslo_concurrency.lockutils [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Lock "6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 865.350131] env[65522]: DEBUG nova.compute.manager [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] No waiting events found dispatching network-vif-plugged-e296c3b2-a6e6-4f3a-8fc8-9a208bb85470 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 865.350131] env[65522]: WARNING nova.compute.manager [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Received unexpected event network-vif-plugged-e296c3b2-a6e6-4f3a-8fc8-9a208bb85470 for instance with vm_state building and task_state spawning. [ 865.350131] env[65522]: DEBUG nova.compute.manager [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Received event network-changed-e296c3b2-a6e6-4f3a-8fc8-9a208bb85470 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 865.350131] env[65522]: DEBUG nova.compute.manager [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Refreshing instance network info cache due to event network-changed-e296c3b2-a6e6-4f3a-8fc8-9a208bb85470. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 865.350131] env[65522]: DEBUG oslo_concurrency.lockutils [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Acquiring lock "refresh_cache-6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.350131] env[65522]: DEBUG oslo_concurrency.lockutils [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Acquired lock "refresh_cache-6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.350131] env[65522]: DEBUG nova.network.neutron [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Refreshing network info cache for port e296c3b2-a6e6-4f3a-8fc8-9a208bb85470 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 865.374131] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114151, 'name': CreateVM_Task, 'duration_secs': 0.456435} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.374340] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 865.374937] env[65522]: WARNING neutronclient.v2_0.client [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.375365] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.375527] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.375880] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 865.376177] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-039d076b-2710-4e15-aafa-41e1cf9d406c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.382618] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 865.382618] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fb4d31-db41-dc91-f418-9677810fb6af" [ 865.382618] env[65522]: _type = "Task" [ 865.382618] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.391711] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fb4d31-db41-dc91-f418-9677810fb6af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.433218] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114152, 'name': CreateVM_Task, 'duration_secs': 0.394677} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.435602] env[65522]: DEBUG nova.network.neutron [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating instance_info_cache with network_info: [{"id": "9446c1bb-b443-4a52-879d-5fa6cb83f0dc", "address": "fa:16:3e:54:30:7c", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9446c1bb-b4", "ovs_interfaceid": "9446c1bb-b443-4a52-879d-5fa6cb83f0dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 865.439671] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 865.440609] env[65522]: DEBUG oslo_concurrency.lockutils [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.446738] env[65522]: DEBUG oslo_vmware.api [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114153, 'name': PowerOffVM_Task, 'duration_secs': 0.268754} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.447067] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 865.447252] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 865.447524] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b6c75ff0-fddd-4b0b-a396-f7edf3db2d94 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.459866] env[65522]: DEBUG nova.compute.manager [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 865.460768] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2952b907-3ab1-489c-968a-07beaa75356a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.526921] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 865.527253] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 865.527744] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Deleting the datastore file [datastore2] 361e54c7-4309-48b3-ad3e-b2ebc783741c {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 865.527744] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-47d92e9a-00a3-4240-a6b5-89107ad39de6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.535456] env[65522]: DEBUG oslo_vmware.api [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for the task: (returnval){ [ 865.535456] env[65522]: value = "task-5114155" [ 865.535456] env[65522]: _type = "Task" [ 865.535456] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.544748] env[65522]: DEBUG oslo_vmware.api [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114155, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.566277] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114150, 'name': ReconfigVM_Task, 'duration_secs': 0.61086} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.568044] env[65522]: DEBUG nova.network.neutron [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Successfully created port: c6221fb0-8c5b-478e-937e-3997556d0aaa {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 865.574321] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 8eda6816-6781-4ad9-9210-11fa623f24b0/8eda6816-6781-4ad9-9210-11fa623f24b0.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.574321] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-59fd748a-d089-4873-b155-063c9b741654 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.580268] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Waiting for the task: (returnval){ [ 865.580268] env[65522]: value = "task-5114156" [ 865.580268] env[65522]: _type = "Task" [ 865.580268] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.591869] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114156, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.616822] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52afb45b-cf37-9943-1f04-581a02483d62, 'name': SearchDatastore_Task, 'duration_secs': 0.014413} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.620705] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecb4d82b-fb28-41b4-bdff-ef74a2d7132a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.623577] env[65522]: DEBUG nova.compute.manager [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 865.632868] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 865.632868] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528bc421-5388-19e0-db79-04b83c3a4f50" [ 865.632868] env[65522]: _type = "Task" [ 865.632868] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.641655] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528bc421-5388-19e0-db79-04b83c3a4f50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.762580] env[65522]: DEBUG nova.compute.manager [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Stashing vm_state: active {{(pid=65522) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 865.856066] env[65522]: WARNING neutronclient.v2_0.client [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.857101] env[65522]: WARNING openstack [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.857884] env[65522]: WARNING openstack [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.899137] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fb4d31-db41-dc91-f418-9677810fb6af, 'name': SearchDatastore_Task, 'duration_secs': 0.011323} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.899666] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.900047] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.900366] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.900725] env[65522]: DEBUG oslo_concurrency.lockutils [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.901094] env[65522]: DEBUG oslo_concurrency.lockutils [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 865.901402] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83d4a2eb-a416-4ad2-b9e5-d2f9f859c8a7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.911112] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 865.911112] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520fe5e7-ae02-66c1-b56a-268219236dc7" [ 865.911112] env[65522]: _type = "Task" [ 865.911112] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.922510] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520fe5e7-ae02-66c1-b56a-268219236dc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.942074] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "refresh_cache-495fa98f-a9db-4214-87cc-d29209d3cb62" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.942364] env[65522]: DEBUG nova.objects.instance [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lazy-loading 'migration_context' on Instance uuid 495fa98f-a9db-4214-87cc-d29209d3cb62 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.050094] env[65522]: DEBUG oslo_vmware.api [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Task: {'id': task-5114155, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166404} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.050402] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 866.050609] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 866.050837] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 866.051040] env[65522]: INFO nova.compute.manager [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 866.051306] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 866.051538] env[65522]: DEBUG nova.compute.manager [-] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 866.051685] env[65522]: DEBUG nova.network.neutron [-] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 866.051989] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.052584] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.052850] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.094016] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114156, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.147620] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528bc421-5388-19e0-db79-04b83c3a4f50, 'name': SearchDatastore_Task, 'duration_secs': 0.020806} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.147915] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 866.148284] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 179a37ed-1176-4082-8c68-1be6abd3fd9d/179a37ed-1176-4082-8c68-1be6abd3fd9d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 866.148607] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 866.148862] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 866.149054] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-78af510b-bd3a-47ab-9cb1-a1ed4812d8c3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.151536] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-550b0fb8-62b9-42ae-8c2c-2fb80691c99f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.165902] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 866.165902] env[65522]: value = "task-5114157" [ 866.165902] env[65522]: _type = "Task" [ 866.165902] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.166592] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 866.166766] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 866.168308] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86f42968-b7dc-4414-a59e-ccbc4c6a8d93 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.184627] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114157, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.185387] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 866.185387] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527a704e-34fc-a3f4-5a2e-82c5820c788e" [ 866.185387] env[65522]: _type = "Task" [ 866.185387] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.197018] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527a704e-34fc-a3f4-5a2e-82c5820c788e, 'name': SearchDatastore_Task, 'duration_secs': 0.011062} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.198256] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-182fabe6-b0a7-4b87-8273-fbdd605ed49f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.208800] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 866.208800] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524f7f16-0654-2d1d-401f-b31310da1bef" [ 866.208800] env[65522]: _type = "Task" [ 866.208800] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.218521] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524f7f16-0654-2d1d-401f-b31310da1bef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.220862] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91cfc02-cd3f-4b10-bdd2-f6c33b212815 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.228459] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fbcfc5a-ae80-4229-985c-cba65a6f1700 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.266743] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd69a4a-0a54-4c83-8cf5-6022950e24d6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.278667] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e659eac-54d3-4c7f-9a14-823d63940cd4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.285527] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 866.294063] env[65522]: DEBUG nova.compute.provider_tree [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.425471] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520fe5e7-ae02-66c1-b56a-268219236dc7, 'name': SearchDatastore_Task, 'duration_secs': 0.011034} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.425823] env[65522]: DEBUG oslo_concurrency.lockutils [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 866.426110] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 866.426484] env[65522]: DEBUG oslo_concurrency.lockutils [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.448171] env[65522]: DEBUG nova.objects.base [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Object Instance<495fa98f-a9db-4214-87cc-d29209d3cb62> lazy-loaded attributes: info_cache,migration_context {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 866.449392] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44658235-4567-4769-b45e-e017522a2f89 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.472956] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3dbc8d21-e9f0-4512-bd5b-cfa447f8b0d1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.483021] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd2874c-03fe-4bee-8b96-325846042c2b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.486323] env[65522]: DEBUG oslo_vmware.api [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 866.486323] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528b8c14-a025-0e54-0063-a9213acbd1cb" [ 866.486323] env[65522]: _type = "Task" [ 866.486323] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.493383] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Doing hard reboot of VM {{(pid=65522) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 866.493977] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-fc26e07e-d5c7-4903-ae73-f2db77fd37cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.502067] env[65522]: DEBUG oslo_vmware.api [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528b8c14-a025-0e54-0063-a9213acbd1cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.505684] env[65522]: DEBUG oslo_vmware.api [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 866.505684] env[65522]: value = "task-5114158" [ 866.505684] env[65522]: _type = "Task" [ 866.505684] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.520505] env[65522]: DEBUG oslo_vmware.api [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114158, 'name': ResetVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.597864] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114156, 'name': Rename_Task, 'duration_secs': 0.670107} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.598270] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 866.598576] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e381daa3-d13d-47f0-b8ed-c2167cca0eda {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.609586] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Waiting for the task: (returnval){ [ 866.609586] env[65522]: value = "task-5114159" [ 866.609586] env[65522]: _type = "Task" [ 866.609586] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.614549] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.622335] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114159, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.634070] env[65522]: DEBUG nova.compute.manager [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 866.664639] env[65522]: DEBUG nova.virt.hardware [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 866.664971] env[65522]: DEBUG nova.virt.hardware [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 866.665177] env[65522]: DEBUG nova.virt.hardware [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 866.665368] env[65522]: DEBUG nova.virt.hardware [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 866.665554] env[65522]: DEBUG nova.virt.hardware [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 866.665706] env[65522]: DEBUG nova.virt.hardware [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 866.665942] env[65522]: DEBUG nova.virt.hardware [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 866.666112] env[65522]: DEBUG nova.virt.hardware [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 866.666313] env[65522]: DEBUG nova.virt.hardware [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 866.666459] env[65522]: DEBUG nova.virt.hardware [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 866.666637] env[65522]: DEBUG nova.virt.hardware [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 866.667735] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf30a2f7-6e41-4004-b945-5998dc9f4c60 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.690550] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114157, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.691625] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd23a64-0f14-437d-85ff-4d1124f01cd8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.720683] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524f7f16-0654-2d1d-401f-b31310da1bef, 'name': SearchDatastore_Task, 'duration_secs': 0.016315} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.720966] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 866.721274] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e/6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 866.721876] env[65522]: DEBUG oslo_concurrency.lockutils [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 866.721876] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 866.722027] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-919bcdba-3ec1-4fd1-8def-11a80d3a41ce {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.724456] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f446731-575d-47d5-8a23-ca25ea10b378 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.732678] env[65522]: WARNING openstack [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.733129] env[65522]: WARNING openstack [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.743412] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 866.743412] env[65522]: value = "task-5114160" [ 866.743412] env[65522]: _type = "Task" [ 866.743412] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.743948] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 866.743948] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 866.744755] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9eb3e07f-29ba-4e35-b189-614e84869e3b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.756751] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 866.756751] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52443b54-ba9d-c2a9-b783-782d36ba102a" [ 866.756751] env[65522]: _type = "Task" [ 866.756751] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.760572] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114160, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.770953] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52443b54-ba9d-c2a9-b783-782d36ba102a, 'name': SearchDatastore_Task, 'duration_secs': 0.011456} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.771799] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91187b40-1b0e-426b-b6e5-0c2b045340e2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.777929] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 866.777929] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5222ff70-a2eb-3f99-66c7-472d0bafb55d" [ 866.777929] env[65522]: _type = "Task" [ 866.777929] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.786592] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5222ff70-a2eb-3f99-66c7-472d0bafb55d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.797906] env[65522]: DEBUG nova.scheduler.client.report [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 866.999572] env[65522]: DEBUG oslo_vmware.api [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528b8c14-a025-0e54-0063-a9213acbd1cb, 'name': SearchDatastore_Task, 'duration_secs': 0.032077} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.999899] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 867.015723] env[65522]: DEBUG oslo_vmware.api [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114158, 'name': ResetVM_Task, 'duration_secs': 0.175073} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.016089] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Did hard reboot of VM {{(pid=65522) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 867.016290] env[65522]: DEBUG nova.compute.manager [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 867.017626] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10d9997-9a24-440a-98ea-f6cbccd82135 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.119534] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114159, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.121027] env[65522]: WARNING neutronclient.v2_0.client [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 867.121531] env[65522]: WARNING openstack [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 867.121894] env[65522]: WARNING openstack [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 867.183891] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114157, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55801} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.184280] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 179a37ed-1176-4082-8c68-1be6abd3fd9d/179a37ed-1176-4082-8c68-1be6abd3fd9d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 867.184564] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 867.184916] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4fc6cd76-32a1-4db1-9884-cd62974d6212 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.195536] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 867.195536] env[65522]: value = "task-5114161" [ 867.195536] env[65522]: _type = "Task" [ 867.195536] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.196709] env[65522]: DEBUG nova.network.neutron [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Successfully updated port: c6221fb0-8c5b-478e-937e-3997556d0aaa {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 867.220462] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114161, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.261164] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114160, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.295613] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5222ff70-a2eb-3f99-66c7-472d0bafb55d, 'name': SearchDatastore_Task, 'duration_secs': 0.245582} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.296411] env[65522]: DEBUG oslo_concurrency.lockutils [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 867.296740] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 26432ad6-610e-49c2-8699-f1e7e128e5b8/26432ad6-610e-49c2-8699-f1e7e128e5b8.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 867.297151] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8fd821bd-fcc1-437a-bcbf-97fa216880ff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.304577] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.697s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 867.305797] env[65522]: DEBUG nova.compute.manager [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 867.311688] env[65522]: DEBUG nova.network.neutron [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Updated VIF entry in instance network info cache for port e296c3b2-a6e6-4f3a-8fc8-9a208bb85470. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 867.312138] env[65522]: DEBUG nova.network.neutron [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Updating instance_info_cache with network_info: [{"id": "e296c3b2-a6e6-4f3a-8fc8-9a208bb85470", "address": "fa:16:3e:31:c2:f0", "network": {"id": "0ab23355-59a7-409b-89f3-089fbd03db5e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-960579569-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "acbde84c090e4c0c919728b809c73d02", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape296c3b2-a6", "ovs_interfaceid": "e296c3b2-a6e6-4f3a-8fc8-9a208bb85470", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 867.315397] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.218s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 867.315692] env[65522]: DEBUG nova.objects.instance [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Lazy-loading 'resources' on Instance uuid 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.317123] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 867.317123] env[65522]: value = "task-5114162" [ 867.317123] env[65522]: _type = "Task" [ 867.317123] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.333064] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114162, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.434623] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Acquiring lock "55ab4771-9908-4640-a142-3cb40c0c6ee5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 867.434623] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Lock "55ab4771-9908-4640-a142-3cb40c0c6ee5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 867.479588] env[65522]: DEBUG nova.network.neutron [-] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 867.532606] env[65522]: DEBUG oslo_concurrency.lockutils [None req-de823ee4-f205-4505-8d9c-e95fc40b4e47 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.637s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 867.623320] env[65522]: DEBUG oslo_vmware.api [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114159, 'name': PowerOnVM_Task, 'duration_secs': 0.777909} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.623594] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.623787] env[65522]: INFO nova.compute.manager [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Took 13.39 seconds to spawn the instance on the hypervisor. [ 867.623969] env[65522]: DEBUG nova.compute.manager [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 867.624829] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32db71e0-d2ec-4e4c-9bf0-e93650debadc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.652289] env[65522]: DEBUG nova.compute.manager [req-47920d99-7505-4168-b4dc-59000e35ea98 req-9fadbe4e-3371-4712-b8dc-b20b709cd8c5 service nova] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Received event network-vif-deleted-0345f514-e2bb-40e4-9692-c753d1dad222 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 867.675997] env[65522]: DEBUG nova.compute.manager [req-f178e7cd-2623-40c4-ba2b-f974337192c3 req-dffc8ea1-689d-453f-9607-8d8ff689936b service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Received event network-vif-plugged-c6221fb0-8c5b-478e-937e-3997556d0aaa {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 867.676899] env[65522]: DEBUG oslo_concurrency.lockutils [req-f178e7cd-2623-40c4-ba2b-f974337192c3 req-dffc8ea1-689d-453f-9607-8d8ff689936b service nova] Acquiring lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 867.676988] env[65522]: DEBUG oslo_concurrency.lockutils [req-f178e7cd-2623-40c4-ba2b-f974337192c3 req-dffc8ea1-689d-453f-9607-8d8ff689936b service nova] Lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 867.677213] env[65522]: DEBUG oslo_concurrency.lockutils [req-f178e7cd-2623-40c4-ba2b-f974337192c3 req-dffc8ea1-689d-453f-9607-8d8ff689936b service nova] Lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 867.677335] env[65522]: DEBUG nova.compute.manager [req-f178e7cd-2623-40c4-ba2b-f974337192c3 req-dffc8ea1-689d-453f-9607-8d8ff689936b service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] No waiting events found dispatching network-vif-plugged-c6221fb0-8c5b-478e-937e-3997556d0aaa {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 867.677493] env[65522]: WARNING nova.compute.manager [req-f178e7cd-2623-40c4-ba2b-f974337192c3 req-dffc8ea1-689d-453f-9607-8d8ff689936b service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Received unexpected event network-vif-plugged-c6221fb0-8c5b-478e-937e-3997556d0aaa for instance with vm_state building and task_state spawning. [ 867.702979] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.703169] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 867.703469] env[65522]: DEBUG nova.network.neutron [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 867.718932] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114161, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.13254} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.719270] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.720259] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0462bd3-c2f5-4e5b-895c-5a60f082c7f9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.751335] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 179a37ed-1176-4082-8c68-1be6abd3fd9d/179a37ed-1176-4082-8c68-1be6abd3fd9d.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.752295] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3cd7d06b-cd70-40fe-ad6c-bc01f8d82102 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.783904] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114160, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.755334} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.786142] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e/6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 867.786277] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 867.786636] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 867.786636] env[65522]: value = "task-5114163" [ 867.786636] env[65522]: _type = "Task" [ 867.786636] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.787137] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2e25c6e4-7fcd-4775-bafb-4ca96438ca4d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.805173] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114163, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.807794] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 867.807794] env[65522]: value = "task-5114164" [ 867.807794] env[65522]: _type = "Task" [ 867.807794] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.825787] env[65522]: DEBUG nova.compute.utils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 867.827467] env[65522]: DEBUG oslo_concurrency.lockutils [req-6894e0c8-4075-4945-b84b-0218ea8e3c70 req-ff8623e2-3808-4d69-ad34-ec600f2cbfc4 service nova] Releasing lock "refresh_cache-6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 867.830222] env[65522]: DEBUG nova.compute.manager [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 867.830222] env[65522]: DEBUG nova.network.neutron [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 867.830222] env[65522]: WARNING neutronclient.v2_0.client [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 867.830833] env[65522]: WARNING neutronclient.v2_0.client [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 867.831682] env[65522]: WARNING openstack [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 867.832465] env[65522]: WARNING openstack [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 867.861981] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114162, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.914293] env[65522]: DEBUG nova.policy [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2109e40451943a3884012e51bf118f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b5c14a856d454048856539895f220c2c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 867.939998] env[65522]: DEBUG nova.compute.manager [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 867.981798] env[65522]: INFO nova.compute.manager [-] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Took 1.93 seconds to deallocate network for instance. [ 868.152211] env[65522]: INFO nova.compute.manager [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Took 48.64 seconds to build instance. [ 868.211070] env[65522]: WARNING openstack [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 868.211070] env[65522]: WARNING openstack [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.242838] env[65522]: DEBUG nova.network.neutron [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Successfully created port: 0750570e-fa7b-4296-a6f5-dec88fcbbab6 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 868.275414] env[65522]: DEBUG nova.network.neutron [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 868.304438] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114163, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.322904] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114164, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117363} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.323237] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 868.324114] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5684f97-570c-4f8f-b381-deb947f1b6a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.330286] env[65522]: DEBUG nova.compute.manager [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 868.358658] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e/6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 868.363217] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ef1b626-d470-4ed3-b8b9-cb92792c1633 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.380338] env[65522]: WARNING openstack [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 868.380696] env[65522]: WARNING openstack [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.395140] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114162, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.652611} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.396691] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 26432ad6-610e-49c2-8699-f1e7e128e5b8/26432ad6-610e-49c2-8699-f1e7e128e5b8.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 868.396925] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 868.397270] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 868.397270] env[65522]: value = "task-5114165" [ 868.397270] env[65522]: _type = "Task" [ 868.397270] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.397469] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-99ad71fb-999b-4e81-8fb9-d0ca4571b3e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.413260] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114165, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.414870] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 868.414870] env[65522]: value = "task-5114166" [ 868.414870] env[65522]: _type = "Task" [ 868.414870] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.420061] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36da415-c8de-4c31-846a-72dbbd77482a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.430975] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32f4e25e-0338-42be-a27a-a24f8d2f2966 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.469551] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Acquiring lock "8eda6816-6781-4ad9-9210-11fa623f24b0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 868.477720] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1299385-8198-427f-93f4-70ac6cc160be {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.487031] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d068ea88-ac8b-4147-9d2c-6c9148988f61 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.491991] env[65522]: DEBUG oslo_concurrency.lockutils [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 868.502876] env[65522]: DEBUG nova.compute.provider_tree [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.505195] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 868.540921] env[65522]: WARNING neutronclient.v2_0.client [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.541719] env[65522]: WARNING openstack [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 868.542416] env[65522]: WARNING openstack [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.661945] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a5fd0652-67fa-45db-9025-04f9b1898d97 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Lock "8eda6816-6781-4ad9-9210-11fa623f24b0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.153s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 868.663557] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Lock "8eda6816-6781-4ad9-9210-11fa623f24b0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.194s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 868.663841] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Acquiring lock "8eda6816-6781-4ad9-9210-11fa623f24b0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 868.664114] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Lock "8eda6816-6781-4ad9-9210-11fa623f24b0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 868.664329] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Lock "8eda6816-6781-4ad9-9210-11fa623f24b0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 868.667429] env[65522]: INFO nova.compute.manager [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Terminating instance [ 868.727579] env[65522]: DEBUG nova.network.neutron [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updating instance_info_cache with network_info: [{"id": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "address": "fa:16:3e:8e:34:0c", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6221fb0-8c", "ovs_interfaceid": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 868.803412] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114163, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.913162] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114165, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.926949] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114166, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.281519} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.927257] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 868.928231] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b23dda0-8207-4d46-970a-1033860779f9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.951454] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 26432ad6-610e-49c2-8699-f1e7e128e5b8/26432ad6-610e-49c2-8699-f1e7e128e5b8.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 868.951814] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6722c3ce-99d9-47bc-aa56-65111d18a767 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.974025] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 868.974025] env[65522]: value = "task-5114167" [ 868.974025] env[65522]: _type = "Task" [ 868.974025] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.985662] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114167, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.007085] env[65522]: DEBUG nova.scheduler.client.report [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 869.176040] env[65522]: DEBUG nova.compute.manager [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 869.176040] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 869.176040] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e85e96-141f-46da-a90c-cea681fcbaa4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.183109] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.183412] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41bc1591-5b0e-4534-9fdb-f74f0a7e8af0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.192874] env[65522]: DEBUG oslo_vmware.api [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Waiting for the task: (returnval){ [ 869.192874] env[65522]: value = "task-5114168" [ 869.192874] env[65522]: _type = "Task" [ 869.192874] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.202981] env[65522]: DEBUG oslo_vmware.api [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114168, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.230965] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 869.231459] env[65522]: DEBUG nova.compute.manager [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Instance network_info: |[{"id": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "address": "fa:16:3e:8e:34:0c", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6221fb0-8c", "ovs_interfaceid": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 869.232010] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:34:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4345ef6-a7c8-4c1c-badf-a0d4f578b61c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6221fb0-8c5b-478e-937e-3997556d0aaa', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 869.243298] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 869.243298] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 869.243298] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7a9c102-e9e4-46dc-81c1-b19ae9844a57 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.265061] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 869.265061] env[65522]: value = "task-5114169" [ 869.265061] env[65522]: _type = "Task" [ 869.265061] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.275514] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114169, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.304679] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114163, 'name': ReconfigVM_Task, 'duration_secs': 1.038624} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.305189] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 179a37ed-1176-4082-8c68-1be6abd3fd9d/179a37ed-1176-4082-8c68-1be6abd3fd9d.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.306156] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5c328a2c-627f-4cac-a392-bc4343baf25d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.315711] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 869.315711] env[65522]: value = "task-5114170" [ 869.315711] env[65522]: _type = "Task" [ 869.315711] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.331868] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114170, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.371021] env[65522]: DEBUG nova.compute.manager [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 869.400342] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 869.400754] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 869.401089] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 869.401332] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 869.401510] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 869.405692] env[65522]: DEBUG nova.virt.hardware [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 869.405962] env[65522]: DEBUG nova.virt.hardware [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 869.406235] env[65522]: DEBUG nova.virt.hardware [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 869.406336] env[65522]: DEBUG nova.virt.hardware [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 869.406483] env[65522]: DEBUG nova.virt.hardware [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 869.406631] env[65522]: DEBUG nova.virt.hardware [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 869.406854] env[65522]: DEBUG nova.virt.hardware [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.407065] env[65522]: DEBUG nova.virt.hardware [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 869.407252] env[65522]: DEBUG nova.virt.hardware [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 869.407418] env[65522]: DEBUG nova.virt.hardware [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 869.407655] env[65522]: DEBUG nova.virt.hardware [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 869.408765] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56acc38a-d6af-4d78-b830-a55eb181e4ab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.413172] env[65522]: INFO nova.compute.manager [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Terminating instance [ 869.430096] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7226665-fc8d-4f0f-be36-8806a0450dd2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.436682] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114165, 'name': ReconfigVM_Task, 'duration_secs': 0.947482} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.437673] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e/6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.438916] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b3f97968-085f-49a6-b8ba-59c05ac4c71f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.459531] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 869.459531] env[65522]: value = "task-5114171" [ 869.459531] env[65522]: _type = "Task" [ 869.459531] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.469702] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114171, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.485562] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114167, 'name': ReconfigVM_Task, 'duration_secs': 0.37426} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.486598] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 26432ad6-610e-49c2-8699-f1e7e128e5b8/26432ad6-610e-49c2-8699-f1e7e128e5b8.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.487215] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e961c82-b1fe-470c-afca-a3bc4f6ca070 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.494939] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 869.494939] env[65522]: value = "task-5114172" [ 869.494939] env[65522]: _type = "Task" [ 869.494939] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.505968] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114172, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.513505] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.198s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 869.518763] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.846s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 869.519292] env[65522]: DEBUG nova.objects.instance [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lazy-loading 'resources' on Instance uuid 97dc8cac-8241-4912-a3ed-689439ef7ff8 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.552187] env[65522]: INFO nova.scheduler.client.report [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Deleted allocations for instance 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20 [ 869.704081] env[65522]: DEBUG oslo_vmware.api [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114168, 'name': PowerOffVM_Task, 'duration_secs': 0.262846} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.704347] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 869.704512] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 869.704787] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c722c1c1-3383-4f2a-bf55-f79b4948ca40 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.772036] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 869.772291] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 869.772445] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Deleting the datastore file [datastore1] 8eda6816-6781-4ad9-9210-11fa623f24b0 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.773277] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8bb6fa86-685e-47de-b9a7-b7fbfca524f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.778479] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114169, 'name': CreateVM_Task, 'duration_secs': 0.376059} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.779035] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 869.779604] env[65522]: WARNING neutronclient.v2_0.client [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 869.780129] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.780346] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 869.780692] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 869.781094] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b0059f7-05bb-4fd7-b127-27fc2b590bcf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.787371] env[65522]: DEBUG oslo_vmware.api [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Waiting for the task: (returnval){ [ 869.787371] env[65522]: value = "task-5114174" [ 869.787371] env[65522]: _type = "Task" [ 869.787371] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.788237] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 869.788237] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52343671-e5ae-9546-c179-a6b674d3da8a" [ 869.788237] env[65522]: _type = "Task" [ 869.788237] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.803676] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52343671-e5ae-9546-c179-a6b674d3da8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.807554] env[65522]: DEBUG oslo_vmware.api [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114174, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.827224] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114170, 'name': Rename_Task, 'duration_secs': 0.201277} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.827882] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 869.827882] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de2a0c94-7b75-42aa-8da3-922260b8a0aa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.835754] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 869.835754] env[65522]: value = "task-5114175" [ 869.835754] env[65522]: _type = "Task" [ 869.835754] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.842611] env[65522]: DEBUG nova.network.neutron [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Successfully updated port: 0750570e-fa7b-4296-a6f5-dec88fcbbab6 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 869.846417] env[65522]: DEBUG nova.compute.manager [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Received event network-changed-cb206964-4331-495a-94ce-c2b95dccc600 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 869.846595] env[65522]: DEBUG nova.compute.manager [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Refreshing instance network info cache due to event network-changed-cb206964-4331-495a-94ce-c2b95dccc600. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 869.846810] env[65522]: DEBUG oslo_concurrency.lockutils [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] Acquiring lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.846944] env[65522]: DEBUG oslo_concurrency.lockutils [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] Acquired lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 869.847109] env[65522]: DEBUG nova.network.neutron [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Refreshing network info cache for port cb206964-4331-495a-94ce-c2b95dccc600 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 869.856643] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114175, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.868981] env[65522]: DEBUG nova.compute.manager [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Received event network-changed-c6221fb0-8c5b-478e-937e-3997556d0aaa {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 869.869277] env[65522]: DEBUG nova.compute.manager [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Refreshing instance network info cache due to event network-changed-c6221fb0-8c5b-478e-937e-3997556d0aaa. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 869.869500] env[65522]: DEBUG oslo_concurrency.lockutils [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] Acquiring lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.870435] env[65522]: DEBUG oslo_concurrency.lockutils [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] Acquired lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 869.870648] env[65522]: DEBUG nova.network.neutron [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Refreshing network info cache for port c6221fb0-8c5b-478e-937e-3997556d0aaa {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 869.920943] env[65522]: DEBUG nova.compute.manager [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 869.921360] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 869.922727] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f3433c-130e-493a-b30e-a346794cfdae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.931385] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.931914] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d6a7f7f2-d9c0-4a30-9017-68a0660e3d8a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.940334] env[65522]: DEBUG oslo_vmware.api [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 869.940334] env[65522]: value = "task-5114176" [ 869.940334] env[65522]: _type = "Task" [ 869.940334] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.952212] env[65522]: DEBUG oslo_vmware.api [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114176, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.971670] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114171, 'name': Rename_Task, 'duration_secs': 0.198274} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.973634] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 869.973634] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e443b672-d46d-4883-abce-23debf6646c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.980883] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 869.980883] env[65522]: value = "task-5114177" [ 869.980883] env[65522]: _type = "Task" [ 869.980883] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.995175] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114177, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.007388] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114172, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.060453] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e05ef846-70c2-4a7b-bd8f-34a96a8c55d8 tempest-VolumesAssistedSnapshotsTest-1945422185 tempest-VolumesAssistedSnapshotsTest-1945422185-project-member] Lock "2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.263s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 870.304580] env[65522]: DEBUG oslo_vmware.api [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Task: {'id': task-5114174, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.26227} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.304895] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.306428] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 870.306688] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 870.306837] env[65522]: INFO nova.compute.manager [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Took 1.13 seconds to destroy the instance on the hypervisor. [ 870.307133] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 870.312870] env[65522]: DEBUG nova.compute.manager [-] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 870.313116] env[65522]: DEBUG nova.network.neutron [-] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 870.314349] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 870.314349] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.314349] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.321810] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52343671-e5ae-9546-c179-a6b674d3da8a, 'name': SearchDatastore_Task, 'duration_secs': 0.014177} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.325485] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 870.325745] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 870.325964] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.326219] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 870.326356] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 870.327055] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be8c51a0-9004-46c4-a644-8e0d370fc76f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.344425] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 870.344618] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 870.346249] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da5391ed-2df4-47d0-9563-3126e2bc8954 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.352997] env[65522]: WARNING neutronclient.v2_0.client [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 870.353691] env[65522]: WARNING openstack [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.354264] env[65522]: WARNING openstack [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.361362] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "refresh_cache-1e7b8237-34ea-479e-b5a7-b1846661d61d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.361535] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "refresh_cache-1e7b8237-34ea-479e-b5a7-b1846661d61d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 870.361708] env[65522]: DEBUG nova.network.neutron [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 870.362960] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114175, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.370345] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 870.370345] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526ccbe7-13df-a871-6893-fde55dffb814" [ 870.370345] env[65522]: _type = "Task" [ 870.370345] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.377846] env[65522]: WARNING neutronclient.v2_0.client [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 870.378234] env[65522]: WARNING openstack [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.378527] env[65522]: WARNING openstack [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.394153] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526ccbe7-13df-a871-6893-fde55dffb814, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.454934] env[65522]: DEBUG oslo_vmware.api [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114176, 'name': PowerOffVM_Task, 'duration_secs': 0.263189} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.455191] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 870.455445] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 870.455680] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a40adc8d-1e00-43e2-9094-3486dfa30bec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.498678] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114177, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.509738] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114172, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.528698] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 870.553895] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 870.555059] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 870.555059] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Deleting the datastore file [datastore2] 77b3ca48-d2c6-4743-a44a-53b9aa84662a {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 870.555059] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e105fb82-538e-45fb-8b0f-7d2cb58a4451 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.568271] env[65522]: DEBUG oslo_vmware.api [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 870.568271] env[65522]: value = "task-5114179" [ 870.568271] env[65522]: _type = "Task" [ 870.568271] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.578594] env[65522]: DEBUG oslo_vmware.api [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114179, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.635915] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b37c7a-32d2-4a32-9c87-32eb1b7161e4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.650086] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8694df1-8805-4765-b8cf-95c347263e03 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.689160] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1039ced2-c90e-4a7b-ad82-0a1499cf47c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.697872] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc88f7ff-ffc1-40ed-ba81-e370abf2eaba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.713445] env[65522]: DEBUG nova.compute.provider_tree [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.858236] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114175, 'name': PowerOnVM_Task, 'duration_secs': 0.684769} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.858703] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 870.860370] env[65522]: INFO nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Took 11.92 seconds to spawn the instance on the hypervisor. [ 870.860696] env[65522]: DEBUG nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 870.862009] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce04223-57dc-4467-9d42-ee43e97a7816 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.869837] env[65522]: WARNING openstack [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.870678] env[65522]: WARNING openstack [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.906758] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526ccbe7-13df-a871-6893-fde55dffb814, 'name': SearchDatastore_Task, 'duration_secs': 0.011466} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.908109] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-970acd3b-3655-4835-9554-5d6e45085411 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.916942] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 870.916942] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c7a10a-3051-6562-fcc6-fe38c7fd4b28" [ 870.916942] env[65522]: _type = "Task" [ 870.916942] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.928517] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c7a10a-3051-6562-fcc6-fe38c7fd4b28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.993080] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114177, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.010586] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114172, 'name': Rename_Task, 'duration_secs': 1.196001} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.011202] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 871.011447] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9edc78f0-8743-4f6e-9adc-5886849e30a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.020281] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 871.020281] env[65522]: value = "task-5114180" [ 871.020281] env[65522]: _type = "Task" [ 871.020281] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.030629] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114180, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.045759] env[65522]: WARNING openstack [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.046370] env[65522]: WARNING openstack [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.075208] env[65522]: WARNING openstack [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.075835] env[65522]: WARNING openstack [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.092162] env[65522]: DEBUG oslo_vmware.api [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114179, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162587} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.092435] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 871.092615] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 871.092784] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 871.092955] env[65522]: INFO nova.compute.manager [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Took 1.17 seconds to destroy the instance on the hypervisor. [ 871.093244] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 871.093470] env[65522]: DEBUG nova.compute.manager [-] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 871.093572] env[65522]: DEBUG nova.network.neutron [-] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 871.093849] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.094370] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.094690] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.144529] env[65522]: DEBUG nova.network.neutron [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 871.221234] env[65522]: DEBUG nova.scheduler.client.report [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 871.406837] env[65522]: INFO nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Took 39.36 seconds to build instance. [ 871.431723] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c7a10a-3051-6562-fcc6-fe38c7fd4b28, 'name': SearchDatastore_Task, 'duration_secs': 0.011237} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.436574] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 871.436911] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 9df1a51a-2811-4486-a4c6-58d618f2ae7d/9df1a51a-2811-4486-a4c6-58d618f2ae7d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 871.438026] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-06e0f7a2-fb13-4e79-8c87-40fb78653631 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.445824] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 871.445824] env[65522]: value = "task-5114181" [ 871.445824] env[65522]: _type = "Task" [ 871.445824] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.461072] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114181, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.480384] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.492184] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114177, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.519174] env[65522]: WARNING openstack [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.519594] env[65522]: WARNING openstack [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.536534] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114180, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.554766] env[65522]: WARNING neutronclient.v2_0.client [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.554766] env[65522]: WARNING openstack [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.554766] env[65522]: WARNING openstack [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.581719] env[65522]: WARNING neutronclient.v2_0.client [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.582502] env[65522]: WARNING openstack [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.582845] env[65522]: WARNING openstack [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.730697] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.211s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.732655] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 22.566s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.732871] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.733666] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 871.733666] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.602s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.733666] env[65522]: DEBUG nova.objects.instance [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Lazy-loading 'resources' on Instance uuid 880390f5-00f1-4dea-9345-29c06289b340 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.735479] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a88428-4440-4593-9567-ba2f60948bac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.748285] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf749cd-d634-425d-9cbb-b236e7e2f740 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.758726] env[65522]: INFO nova.scheduler.client.report [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Deleted allocations for instance 97dc8cac-8241-4912-a3ed-689439ef7ff8 [ 871.774679] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5148c3bf-55c1-4584-95f1-1813169efc06 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.788234] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ecd7019-0dba-4447-b0dc-82b8cf95fe3d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.826607] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=176625MB free_disk=94GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 871.826814] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 871.831550] env[65522]: DEBUG nova.network.neutron [-] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 871.909504] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "179a37ed-1176-4082-8c68-1be6abd3fd9d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.875s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.964498] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114181, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.980778] env[65522]: WARNING neutronclient.v2_0.client [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.981743] env[65522]: WARNING openstack [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.982115] env[65522]: WARNING openstack [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 872.010174] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114177, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.042779] env[65522]: DEBUG oslo_vmware.api [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114180, 'name': PowerOnVM_Task, 'duration_secs': 0.520994} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.043960] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 872.043960] env[65522]: DEBUG nova.compute.manager [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 872.044952] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993af180-a9d4-414d-9a4f-f3895dd10e00 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.078825] env[65522]: DEBUG nova.network.neutron [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Updated VIF entry in instance network info cache for port cb206964-4331-495a-94ce-c2b95dccc600. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 872.079993] env[65522]: DEBUG nova.network.neutron [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Updating instance_info_cache with network_info: [{"id": "cb206964-4331-495a-94ce-c2b95dccc600", "address": "fa:16:3e:5b:a0:80", "network": {"id": "922eeeeb-e124-4ae5-97a4-0e995b16f7ec", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-437359683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37fc01876567476f9b93d765b2cfddc8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccf76700-491b-4462-ab19-e6d3a9ff87ac", "external-id": "nsx-vlan-transportzone-956", "segmentation_id": 956, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb206964-43", "ovs_interfaceid": "cb206964-4331-495a-94ce-c2b95dccc600", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 872.167430] env[65522]: DEBUG nova.network.neutron [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updated VIF entry in instance network info cache for port c6221fb0-8c5b-478e-937e-3997556d0aaa. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 872.167430] env[65522]: DEBUG nova.network.neutron [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updating instance_info_cache with network_info: [{"id": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "address": "fa:16:3e:8e:34:0c", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6221fb0-8c", "ovs_interfaceid": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 872.272732] env[65522]: DEBUG nova.network.neutron [-] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 872.282788] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b2b92ef-9fd4-4ce3-becd-79015af22f53 tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "97dc8cac-8241-4912-a3ed-689439ef7ff8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.434s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 872.304391] env[65522]: DEBUG nova.network.neutron [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Updating instance_info_cache with network_info: [{"id": "0750570e-fa7b-4296-a6f5-dec88fcbbab6", "address": "fa:16:3e:40:ce:37", "network": {"id": "dc9d8462-3539-4cc5-b21a-e1136a8df8e6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1580876213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b5c14a856d454048856539895f220c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0750570e-fa", "ovs_interfaceid": "0750570e-fa7b-4296-a6f5-dec88fcbbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 872.334389] env[65522]: INFO nova.compute.manager [-] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Took 2.02 seconds to deallocate network for instance. [ 872.465194] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114181, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.509046] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114177, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.576765] env[65522]: DEBUG oslo_concurrency.lockutils [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 872.584515] env[65522]: DEBUG oslo_concurrency.lockutils [req-386ca6f9-4d66-45c7-bee4-1cbd7655e87c req-f9966ba5-9511-410b-8f7f-2857f1b977a1 service nova] Releasing lock "refresh_cache-77b3ca48-d2c6-4743-a44a-53b9aa84662a" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 872.672943] env[65522]: DEBUG oslo_concurrency.lockutils [req-8a2f4949-7937-44e9-9639-71394f819729 req-3d660c2a-eda8-4285-85eb-aeb434933ec0 service nova] Releasing lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 872.775739] env[65522]: INFO nova.compute.manager [-] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Took 1.68 seconds to deallocate network for instance. [ 872.806558] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "refresh_cache-1e7b8237-34ea-479e-b5a7-b1846661d61d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 872.809021] env[65522]: DEBUG nova.compute.manager [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Instance network_info: |[{"id": "0750570e-fa7b-4296-a6f5-dec88fcbbab6", "address": "fa:16:3e:40:ce:37", "network": {"id": "dc9d8462-3539-4cc5-b21a-e1136a8df8e6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1580876213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b5c14a856d454048856539895f220c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0750570e-fa", "ovs_interfaceid": "0750570e-fa7b-4296-a6f5-dec88fcbbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 872.809021] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:ce:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0750570e-fa7b-4296-a6f5-dec88fcbbab6', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.817588] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 872.818213] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.818606] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3db18599-fd89-4712-9f23-2b860d38f38a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.844402] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 872.855614] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.855614] env[65522]: value = "task-5114182" [ 872.855614] env[65522]: _type = "Task" [ 872.855614] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.869836] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114182, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.937711] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09222c10-a4c8-4a01-be46-c048d1e047c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.949637] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d0e006b-42ae-41cb-89cd-7f497836bc43 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.967376] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114181, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.022553} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.999089] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 9df1a51a-2811-4486-a4c6-58d618f2ae7d/9df1a51a-2811-4486-a4c6-58d618f2ae7d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 872.999482] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 873.000955] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3b6777fb-ed9b-4302-9ee9-4e60cfe72ac5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.013284] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f56b17-30d5-40a7-9320-2f46a8902f61 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.020023] env[65522]: DEBUG nova.compute.manager [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Received event network-vif-plugged-0750570e-fa7b-4296-a6f5-dec88fcbbab6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 873.020023] env[65522]: DEBUG oslo_concurrency.lockutils [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] Acquiring lock "1e7b8237-34ea-479e-b5a7-b1846661d61d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.020023] env[65522]: DEBUG oslo_concurrency.lockutils [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] Lock "1e7b8237-34ea-479e-b5a7-b1846661d61d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 873.020023] env[65522]: DEBUG oslo_concurrency.lockutils [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] Lock "1e7b8237-34ea-479e-b5a7-b1846661d61d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 873.020023] env[65522]: DEBUG nova.compute.manager [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] No waiting events found dispatching network-vif-plugged-0750570e-fa7b-4296-a6f5-dec88fcbbab6 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 873.020023] env[65522]: WARNING nova.compute.manager [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Received unexpected event network-vif-plugged-0750570e-fa7b-4296-a6f5-dec88fcbbab6 for instance with vm_state building and task_state spawning. [ 873.020023] env[65522]: DEBUG nova.compute.manager [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Received event network-changed-0750570e-fa7b-4296-a6f5-dec88fcbbab6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 873.020023] env[65522]: DEBUG nova.compute.manager [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Refreshing instance network info cache due to event network-changed-0750570e-fa7b-4296-a6f5-dec88fcbbab6. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 873.020023] env[65522]: DEBUG oslo_concurrency.lockutils [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] Acquiring lock "refresh_cache-1e7b8237-34ea-479e-b5a7-b1846661d61d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.020023] env[65522]: DEBUG oslo_concurrency.lockutils [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] Acquired lock "refresh_cache-1e7b8237-34ea-479e-b5a7-b1846661d61d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 873.020715] env[65522]: DEBUG nova.network.neutron [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Refreshing network info cache for port 0750570e-fa7b-4296-a6f5-dec88fcbbab6 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 873.033067] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114177, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.038902] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 873.038902] env[65522]: value = "task-5114183" [ 873.038902] env[65522]: _type = "Task" [ 873.038902] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.043204] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ffc3fbb-bc9d-46b2-a573-ff4817a6a8a0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.049655] env[65522]: DEBUG nova.compute.manager [req-e6bead08-89a6-4800-a87b-e6a0d94f7ca5 req-554be97a-4feb-4160-ba91-e19d39077896 service nova] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Received event network-vif-deleted-30b8dee9-9bde-4601-af20-2015c5e9526d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 873.049895] env[65522]: DEBUG nova.compute.manager [req-e6bead08-89a6-4800-a87b-e6a0d94f7ca5 req-554be97a-4feb-4160-ba91-e19d39077896 service nova] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Received event network-vif-deleted-cb206964-4331-495a-94ce-c2b95dccc600 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 873.075456] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "9c39b722-56ff-44fa-8f66-3e3432645a68" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.075972] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "9c39b722-56ff-44fa-8f66-3e3432645a68" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 873.077368] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114183, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.077779] env[65522]: DEBUG nova.compute.provider_tree [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.291233] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.309483] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.309721] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 873.369706] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114182, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.511385] env[65522]: DEBUG oslo_vmware.api [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114177, 'name': PowerOnVM_Task, 'duration_secs': 3.133368} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.511843] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 873.511934] env[65522]: INFO nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Took 11.78 seconds to spawn the instance on the hypervisor. [ 873.512071] env[65522]: DEBUG nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 873.513041] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-616437a0-ef5e-43df-9af1-c289d33ac202 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.524824] env[65522]: WARNING neutronclient.v2_0.client [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 873.525748] env[65522]: WARNING openstack [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 873.526142] env[65522]: WARNING openstack [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 873.563726] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114183, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.184232} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.563726] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 873.563726] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09785318-cbd9-4d5d-8243-0139fc14c20b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.583510] env[65522]: DEBUG nova.compute.manager [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 873.596433] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 9df1a51a-2811-4486-a4c6-58d618f2ae7d/9df1a51a-2811-4486-a4c6-58d618f2ae7d.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.601057] env[65522]: DEBUG nova.scheduler.client.report [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 873.604853] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02bb969f-4f44-4c74-bfe2-a757ea244b2a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.631743] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 873.631743] env[65522]: value = "task-5114184" [ 873.631743] env[65522]: _type = "Task" [ 873.631743] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.643037] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114184, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.677507] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.678177] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 873.678485] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.678765] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 873.678959] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 873.681877] env[65522]: INFO nova.compute.manager [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Terminating instance [ 873.720409] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquiring lock "26432ad6-610e-49c2-8699-f1e7e128e5b8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.720672] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Lock "26432ad6-610e-49c2-8699-f1e7e128e5b8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 873.720927] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquiring lock "26432ad6-610e-49c2-8699-f1e7e128e5b8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.721664] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Lock "26432ad6-610e-49c2-8699-f1e7e128e5b8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 873.721861] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Lock "26432ad6-610e-49c2-8699-f1e7e128e5b8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 873.729809] env[65522]: INFO nova.compute.manager [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Terminating instance [ 873.779506] env[65522]: WARNING openstack [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 873.780268] env[65522]: WARNING openstack [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 873.814391] env[65522]: DEBUG nova.compute.utils [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 873.871247] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114182, 'name': CreateVM_Task, 'duration_secs': 0.551527} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.871575] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 873.872388] env[65522]: WARNING neutronclient.v2_0.client [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 873.872973] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.873266] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 873.873780] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 873.874207] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34d8be3e-0486-4843-bc01-903e63163d27 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.881781] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 873.881781] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522287cc-9742-e13e-8fba-f9b4e6a751ef" [ 873.881781] env[65522]: _type = "Task" [ 873.881781] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.892519] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522287cc-9742-e13e-8fba-f9b4e6a751ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.926154] env[65522]: WARNING neutronclient.v2_0.client [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 873.926546] env[65522]: WARNING openstack [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 873.926900] env[65522]: WARNING openstack [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.044025] env[65522]: INFO nova.compute.manager [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Took 41.97 seconds to build instance. [ 874.086444] env[65522]: DEBUG nova.network.neutron [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Updated VIF entry in instance network info cache for port 0750570e-fa7b-4296-a6f5-dec88fcbbab6. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 874.086951] env[65522]: DEBUG nova.network.neutron [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Updating instance_info_cache with network_info: [{"id": "0750570e-fa7b-4296-a6f5-dec88fcbbab6", "address": "fa:16:3e:40:ce:37", "network": {"id": "dc9d8462-3539-4cc5-b21a-e1136a8df8e6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1580876213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b5c14a856d454048856539895f220c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0750570e-fa", "ovs_interfaceid": "0750570e-fa7b-4296-a6f5-dec88fcbbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 874.124372] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.124640] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.391s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.127139] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.589s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.129259] env[65522]: INFO nova.compute.claims [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.147941] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114184, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.148912] env[65522]: INFO nova.scheduler.client.report [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Deleted allocations for instance 880390f5-00f1-4dea-9345-29c06289b340 [ 874.189777] env[65522]: DEBUG nova.compute.manager [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 874.190061] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.191951] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bdecad2-7f63-4f14-9fdf-5a4584d195cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.203931] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 874.204607] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33f72fac-352a-4c62-a79a-4e76754ff101 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.213142] env[65522]: DEBUG oslo_vmware.api [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 874.213142] env[65522]: value = "task-5114185" [ 874.213142] env[65522]: _type = "Task" [ 874.213142] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.231847] env[65522]: DEBUG oslo_vmware.api [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5114185, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.234951] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquiring lock "refresh_cache-26432ad6-610e-49c2-8699-f1e7e128e5b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.235391] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquired lock "refresh_cache-26432ad6-610e-49c2-8699-f1e7e128e5b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 874.235777] env[65522]: DEBUG nova.network.neutron [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 874.321464] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.395821] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522287cc-9742-e13e-8fba-f9b4e6a751ef, 'name': SearchDatastore_Task, 'duration_secs': 0.014562} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.396248] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 874.396522] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.396836] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.397029] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 874.397272] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.397642] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-192750ea-92d4-4152-8db1-8c0ae1ac95e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.413365] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.413757] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 874.415666] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c712ed3-69a8-4d71-9576-b789a2a6dfed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.425304] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 874.425304] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ce1dc1-5b21-c8d5-91fc-7ff2006fc6d6" [ 874.425304] env[65522]: _type = "Task" [ 874.425304] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.436558] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ce1dc1-5b21-c8d5-91fc-7ff2006fc6d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.546863] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fc9985ce-208c-48d0-a86f-9c4a649a44f4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.484s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.590562] env[65522]: DEBUG oslo_concurrency.lockutils [req-febe4d8d-3f7b-4811-a5b5-9589a16f54f1 req-8ee11a58-9236-47fb-a1c8-cb393b6a2993 service nova] Releasing lock "refresh_cache-1e7b8237-34ea-479e-b5a7-b1846661d61d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 874.647222] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114184, 'name': ReconfigVM_Task, 'duration_secs': 0.627422} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.647222] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 9df1a51a-2811-4486-a4c6-58d618f2ae7d/9df1a51a-2811-4486-a4c6-58d618f2ae7d.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.648095] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-66bc3696-a061-4736-bb60-b5546e60522a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.660948] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 874.660948] env[65522]: value = "task-5114186" [ 874.660948] env[65522]: _type = "Task" [ 874.660948] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.661517] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19067be1-8867-48c8-ab50-a11af7f04124 tempest-ServersNegativeTestMultiTenantJSON-1777378102 tempest-ServersNegativeTestMultiTenantJSON-1777378102-project-member] Lock "880390f5-00f1-4dea-9345-29c06289b340" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.763s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.679256] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114186, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.724777] env[65522]: DEBUG oslo_vmware.api [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5114185, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.739238] env[65522]: WARNING neutronclient.v2_0.client [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 874.740064] env[65522]: WARNING openstack [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.740421] env[65522]: WARNING openstack [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.941201] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ce1dc1-5b21-c8d5-91fc-7ff2006fc6d6, 'name': SearchDatastore_Task, 'duration_secs': 0.027403} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.941266] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-478faa77-cf2a-40ff-b60c-5a9443f665fc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.947990] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 874.947990] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52025a67-2ecd-4323-fe19-1d103f211da4" [ 874.947990] env[65522]: _type = "Task" [ 874.947990] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.958690] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52025a67-2ecd-4323-fe19-1d103f211da4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.049346] env[65522]: DEBUG nova.network.neutron [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 875.173810] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114186, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.226835] env[65522]: DEBUG oslo_vmware.api [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5114185, 'name': PowerOffVM_Task, 'duration_secs': 0.937465} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.227233] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 875.227329] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 875.231214] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-52b0b14f-d3a1-4fac-a77e-8d304729de43 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.322423] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 875.322618] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 875.323080] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Deleting the datastore file [datastore1] e5d2cc50-4ee6-4e64-9b52-888968a717ca {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 875.323080] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-538d73b4-c0e8-4760-8187-64c54b2020e4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.335785] env[65522]: DEBUG oslo_vmware.api [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for the task: (returnval){ [ 875.335785] env[65522]: value = "task-5114188" [ 875.335785] env[65522]: _type = "Task" [ 875.335785] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.347417] env[65522]: DEBUG oslo_vmware.api [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5114188, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.412299] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 875.412299] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 875.412299] env[65522]: INFO nova.compute.manager [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Attaching volume 33425e73-8f97-4928-afed-0733d11034fb to /dev/sdb [ 875.462534] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52025a67-2ecd-4323-fe19-1d103f211da4, 'name': SearchDatastore_Task, 'duration_secs': 0.013354} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.467299] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 875.467600] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 1e7b8237-34ea-479e-b5a7-b1846661d61d/1e7b8237-34ea-479e-b5a7-b1846661d61d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 875.468522] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-126b0d26-854e-46c7-9d9e-2eecacfa0438 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.471604] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e269b13d-b914-4746-be06-bdf51f5892c5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.483817] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92646a6e-e2b3-4551-bc12-17bb21b625ac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.487466] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 875.487466] env[65522]: value = "task-5114189" [ 875.487466] env[65522]: _type = "Task" [ 875.487466] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.504756] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114189, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.511699] env[65522]: DEBUG nova.virt.block_device [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Updating existing volume attachment record: 17605043-e0a2-4d1e-a98c-d214b86e71a8 {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 875.676104] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114186, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.759535] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1b195c-c4b6-4913-9370-d75f15d484f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.766773] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f028679f-a409-4ec8-8665-50402f1a0eeb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.776448] env[65522]: DEBUG nova.network.neutron [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 875.822019] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6dee64-73f8-4578-8cb3-0db898b6b44d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.833475] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce811ed-21ae-44cc-ac4c-bba579604edf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.855431] env[65522]: DEBUG nova.compute.provider_tree [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.860825] env[65522]: DEBUG oslo_vmware.api [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Task: {'id': task-5114188, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.316338} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.860825] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.861526] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.861526] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.861723] env[65522]: INFO nova.compute.manager [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Took 1.67 seconds to destroy the instance on the hypervisor. [ 875.861898] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 875.862170] env[65522]: DEBUG nova.compute.manager [-] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 875.862286] env[65522]: DEBUG nova.network.neutron [-] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 875.862583] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 875.863248] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.863541] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 876.011356] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114189, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.115820] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 876.180684] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114186, 'name': Rename_Task, 'duration_secs': 1.166315} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.181147] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.181515] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b3f935c-f151-4e40-84a0-f7788337cc37 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.190474] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 876.190474] env[65522]: value = "task-5114193" [ 876.190474] env[65522]: _type = "Task" [ 876.190474] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.039630] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Releasing lock "refresh_cache-26432ad6-610e-49c2-8699-f1e7e128e5b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 877.039630] env[65522]: DEBUG nova.compute.manager [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 877.039630] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 877.040258] env[65522]: DEBUG nova.scheduler.client.report [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 877.046630] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114193, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.046630] env[65522]: WARNING oslo_vmware.common.loopingcall [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] task run outlasted interval by 0.35385599999999995 sec [ 877.046901] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84775483-b259-490b-8afe-24ad51426ccd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.067080] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 877.067312] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114189, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.632368} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.067750] env[65522]: DEBUG oslo_vmware.api [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114193, 'name': PowerOnVM_Task, 'duration_secs': 0.557366} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.067750] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f40c8618-1ebd-4000-aba8-4f170b565714 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.072039] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 1e7b8237-34ea-479e-b5a7-b1846661d61d/1e7b8237-34ea-479e-b5a7-b1846661d61d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 877.072039] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 877.072039] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.072039] env[65522]: INFO nova.compute.manager [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Took 10.44 seconds to spawn the instance on the hypervisor. [ 877.072039] env[65522]: DEBUG nova.compute.manager [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 877.072039] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a79dc16a-e0f7-4c59-8a07-4a9ea2274b0a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.073497] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d63859e5-476d-4149-8ef2-76af4676d3cb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.086861] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 877.086861] env[65522]: value = "task-5114195" [ 877.086861] env[65522]: _type = "Task" [ 877.086861] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.089021] env[65522]: DEBUG oslo_vmware.api [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 877.089021] env[65522]: value = "task-5114194" [ 877.089021] env[65522]: _type = "Task" [ 877.089021] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.102694] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114195, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.106607] env[65522]: DEBUG oslo_vmware.api [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114194, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.546580] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.419s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 877.547126] env[65522]: DEBUG nova.compute.manager [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 877.551185] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.146s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 877.555733] env[65522]: INFO nova.compute.claims [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.561579] env[65522]: DEBUG nova.network.neutron [-] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 877.609373] env[65522]: INFO nova.compute.manager [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Took 34.23 seconds to build instance. [ 877.617619] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114195, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.200543} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.621889] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 877.622384] env[65522]: DEBUG oslo_vmware.api [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114194, 'name': PowerOffVM_Task, 'duration_secs': 0.187283} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.623690] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27f3b1d-0d3f-48f1-9655-77175cc9868b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.626394] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 877.626612] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 877.627793] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80c376f3-35ee-4623-a0ec-7d52675b1356 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.656334] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 1e7b8237-34ea-479e-b5a7-b1846661d61d/1e7b8237-34ea-479e-b5a7-b1846661d61d.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.658402] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-afbb54e9-d18d-4593-9775-d822b3c67090 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.680585] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 877.680585] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 877.680585] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Deleting the datastore file [datastore1] 26432ad6-610e-49c2-8699-f1e7e128e5b8 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 877.680585] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e607fcf-8a21-466e-9172-28f9e23d1541 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.690747] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 877.690747] env[65522]: value = "task-5114197" [ 877.690747] env[65522]: _type = "Task" [ 877.690747] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.697719] env[65522]: DEBUG oslo_vmware.api [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for the task: (returnval){ [ 877.697719] env[65522]: value = "task-5114198" [ 877.697719] env[65522]: _type = "Task" [ 877.697719] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.707931] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114197, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.714468] env[65522]: DEBUG oslo_vmware.api [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114198, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.818844] env[65522]: DEBUG nova.compute.manager [req-6f2f805f-86da-4bf2-b0e7-cc1fbef80ffa req-0f8b25f7-c281-4cc1-8173-1074cff6b334 service nova] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Received event network-vif-deleted-3f4547a2-4a63-4803-a79f-eda70e48dd48 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 878.063088] env[65522]: DEBUG nova.compute.utils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 878.068201] env[65522]: DEBUG nova.compute.manager [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 878.068201] env[65522]: DEBUG nova.network.neutron [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 878.068307] env[65522]: WARNING neutronclient.v2_0.client [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.068599] env[65522]: WARNING neutronclient.v2_0.client [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.069194] env[65522]: WARNING openstack [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.069608] env[65522]: WARNING openstack [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.077518] env[65522]: INFO nova.compute.manager [-] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Took 2.22 seconds to deallocate network for instance. [ 878.113653] env[65522]: DEBUG oslo_concurrency.lockutils [None req-062b02c7-8a46-4c66-bde3-486954361ef4 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.744s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 878.129615] env[65522]: DEBUG nova.policy [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '792013a276474561bc625ac0e00cdb82', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2bd38f300424db78568e0b8d9982f27', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 878.179943] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "179a37ed-1176-4082-8c68-1be6abd3fd9d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.181041] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "179a37ed-1176-4082-8c68-1be6abd3fd9d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.181966] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "179a37ed-1176-4082-8c68-1be6abd3fd9d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.181966] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "179a37ed-1176-4082-8c68-1be6abd3fd9d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.182212] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "179a37ed-1176-4082-8c68-1be6abd3fd9d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 878.184906] env[65522]: INFO nova.compute.manager [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Terminating instance [ 878.222769] env[65522]: DEBUG oslo_vmware.api [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Task: {'id': task-5114198, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106027} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.223936] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114197, 'name': ReconfigVM_Task, 'duration_secs': 0.38409} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.224578] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 878.225506] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 878.225832] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 878.227469] env[65522]: INFO nova.compute.manager [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Took 1.19 seconds to destroy the instance on the hypervisor. [ 878.227469] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 878.229178] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 1e7b8237-34ea-479e-b5a7-b1846661d61d/1e7b8237-34ea-479e-b5a7-b1846661d61d.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.229178] env[65522]: DEBUG nova.compute.manager [-] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 878.229178] env[65522]: DEBUG nova.network.neutron [-] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 878.230040] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.230558] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.231136] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.243328] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-74a39649-0324-4650-b3d5-4e38b7ebc76d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.253671] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 878.253671] env[65522]: value = "task-5114200" [ 878.253671] env[65522]: _type = "Task" [ 878.253671] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.266913] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114200, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.514073] env[65522]: DEBUG nova.network.neutron [-] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 878.514073] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.549386] env[65522]: DEBUG nova.network.neutron [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Successfully created port: eba45f53-1f29-46b1-8285-33b5aacbcc43 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 878.566146] env[65522]: DEBUG nova.compute.manager [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 878.586826] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.655629] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.655924] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.656166] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.656366] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.656535] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 878.658980] env[65522]: INFO nova.compute.manager [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Terminating instance [ 878.692318] env[65522]: DEBUG nova.compute.manager [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 878.692318] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 878.696402] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659f4a9c-a546-479e-8c2b-215837201558 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.714047] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 878.716971] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f542ea4-b26e-43d6-931b-89b23c1839cd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.725866] env[65522]: DEBUG oslo_vmware.api [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 878.725866] env[65522]: value = "task-5114201" [ 878.725866] env[65522]: _type = "Task" [ 878.725866] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.737791] env[65522]: DEBUG oslo_vmware.api [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114201, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.769620] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114200, 'name': Rename_Task, 'duration_secs': 0.21591} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.769620] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 878.769620] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6b7a73e-f68d-4a40-9706-1758728f9031 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.779462] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 878.779462] env[65522]: value = "task-5114202" [ 878.779462] env[65522]: _type = "Task" [ 878.779462] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.793254] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114202, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.015547] env[65522]: DEBUG nova.network.neutron [-] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 879.155518] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4b9e7c-b63d-4062-bece-75d000e2a989 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.165363] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32eb97d3-9dec-4a82-9024-8f9fa972fc44 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.169540] env[65522]: DEBUG nova.compute.manager [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 879.169771] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 879.171082] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab402d23-81a0-4a67-b7ab-5df4f767aeab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.179616] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 879.205735] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80e8c0d3-dcbc-478a-bfa9-ff5f4c866977 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.208976] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0007cd4-b894-4b09-8245-36440bed7ec6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.221964] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dbbb167-567d-4192-aade-6291f22df724 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.227211] env[65522]: DEBUG oslo_vmware.api [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 879.227211] env[65522]: value = "task-5114203" [ 879.227211] env[65522]: _type = "Task" [ 879.227211] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.244755] env[65522]: DEBUG nova.compute.provider_tree [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.254062] env[65522]: DEBUG oslo_vmware.api [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114203, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.257409] env[65522]: DEBUG oslo_vmware.api [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114201, 'name': PowerOffVM_Task, 'duration_secs': 0.274553} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.257409] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 879.258544] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 879.258544] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b9cad0e2-8119-4008-96e4-f93868944ef0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.294994] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114202, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.337378] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 879.337849] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 879.338257] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Deleting the datastore file [datastore1] 179a37ed-1176-4082-8c68-1be6abd3fd9d {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 879.338658] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-238695ea-9c13-45b3-9339-c9b9d9a2f81b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.348343] env[65522]: DEBUG oslo_vmware.api [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 879.348343] env[65522]: value = "task-5114205" [ 879.348343] env[65522]: _type = "Task" [ 879.348343] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.359023] env[65522]: DEBUG oslo_vmware.api [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114205, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.522139] env[65522]: INFO nova.compute.manager [-] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Took 1.29 seconds to deallocate network for instance. [ 879.586236] env[65522]: DEBUG nova.compute.manager [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 879.617684] env[65522]: DEBUG nova.virt.hardware [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 879.617944] env[65522]: DEBUG nova.virt.hardware [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 879.618151] env[65522]: DEBUG nova.virt.hardware [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 879.618302] env[65522]: DEBUG nova.virt.hardware [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 879.618443] env[65522]: DEBUG nova.virt.hardware [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 879.618584] env[65522]: DEBUG nova.virt.hardware [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 879.618786] env[65522]: DEBUG nova.virt.hardware [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.618939] env[65522]: DEBUG nova.virt.hardware [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 879.619125] env[65522]: DEBUG nova.virt.hardware [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 879.619287] env[65522]: DEBUG nova.virt.hardware [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 879.619453] env[65522]: DEBUG nova.virt.hardware [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 879.620385] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe895cc-1f8e-49fc-bf12-fb1bbfe40de2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.629239] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535b5f1e-6ec7-4be3-a8c5-ad2041b003b5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.739293] env[65522]: DEBUG oslo_vmware.api [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114203, 'name': PowerOffVM_Task, 'duration_secs': 0.2436} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.739613] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 879.739761] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 879.740185] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0582bc36-b22a-49ff-b264-bf6cd41a66b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.758396] env[65522]: DEBUG nova.scheduler.client.report [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 879.793735] env[65522]: DEBUG oslo_vmware.api [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114202, 'name': PowerOnVM_Task, 'duration_secs': 0.688084} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.793978] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 879.794194] env[65522]: INFO nova.compute.manager [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Took 10.42 seconds to spawn the instance on the hypervisor. [ 879.794366] env[65522]: DEBUG nova.compute.manager [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 879.795291] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abfd5e95-00b1-4793-a843-020c6cb316be {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.821359] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 879.821359] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 879.821359] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Deleting the datastore file [datastore1] 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 879.821570] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5491155-067d-4b9c-89fc-0494a9b71ff1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.830968] env[65522]: DEBUG oslo_vmware.api [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for the task: (returnval){ [ 879.830968] env[65522]: value = "task-5114207" [ 879.830968] env[65522]: _type = "Task" [ 879.830968] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.845579] env[65522]: DEBUG oslo_vmware.api [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114207, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.863695] env[65522]: DEBUG oslo_vmware.api [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114205, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167602} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.863695] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 879.863695] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 879.863695] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 879.863695] env[65522]: INFO nova.compute.manager [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Took 1.17 seconds to destroy the instance on the hypervisor. [ 879.863695] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 879.864570] env[65522]: DEBUG nova.compute.manager [-] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 879.864570] env[65522]: DEBUG nova.network.neutron [-] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 879.864570] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 879.865375] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 879.865834] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 879.982096] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 880.029016] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 880.070629] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 880.070917] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994827', 'volume_id': '33425e73-8f97-4928-afed-0733d11034fb', 'name': 'volume-33425e73-8f97-4928-afed-0733d11034fb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1eebb320-30e2-4d8b-b1fd-19e31fb35b77', 'attached_at': '', 'detached_at': '', 'volume_id': '33425e73-8f97-4928-afed-0733d11034fb', 'serial': '33425e73-8f97-4928-afed-0733d11034fb'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 880.072259] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f346d517-e275-4c36-aacb-6d54c047c84b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.092182] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4406b7e5-cf1c-4156-9154-4309a300e6e4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.117571] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] volume-33425e73-8f97-4928-afed-0733d11034fb/volume-33425e73-8f97-4928-afed-0733d11034fb.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.117986] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ca939ab-c9b4-47cf-9e67-c7205e9bc5c9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.137528] env[65522]: DEBUG oslo_vmware.api [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 880.137528] env[65522]: value = "task-5114208" [ 880.137528] env[65522]: _type = "Task" [ 880.137528] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.147073] env[65522]: DEBUG oslo_vmware.api [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114208, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.268057] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.716s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.268057] env[65522]: DEBUG nova.compute.manager [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 880.269875] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.332s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 880.271454] env[65522]: INFO nova.compute.claims [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.281616] env[65522]: DEBUG nova.compute.manager [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Received event network-changed-c6221fb0-8c5b-478e-937e-3997556d0aaa {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 880.281852] env[65522]: DEBUG nova.compute.manager [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Refreshing instance network info cache due to event network-changed-c6221fb0-8c5b-478e-937e-3997556d0aaa. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 880.282148] env[65522]: DEBUG oslo_concurrency.lockutils [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] Acquiring lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.282389] env[65522]: DEBUG oslo_concurrency.lockutils [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] Acquired lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 880.282796] env[65522]: DEBUG nova.network.neutron [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Refreshing network info cache for port c6221fb0-8c5b-478e-937e-3997556d0aaa {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 880.322448] env[65522]: INFO nova.compute.manager [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Took 35.72 seconds to build instance. [ 880.346024] env[65522]: DEBUG oslo_vmware.api [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Task: {'id': task-5114207, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17587} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.346024] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.346024] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 880.346024] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 880.346326] env[65522]: INFO nova.compute.manager [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Took 1.18 seconds to destroy the instance on the hypervisor. [ 880.346585] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 880.346784] env[65522]: DEBUG nova.compute.manager [-] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 880.346884] env[65522]: DEBUG nova.network.neutron [-] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 880.347152] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 880.347689] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 880.348119] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 880.393161] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 880.437678] env[65522]: DEBUG nova.network.neutron [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Successfully updated port: eba45f53-1f29-46b1-8285-33b5aacbcc43 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 880.543165] env[65522]: DEBUG nova.compute.manager [req-792b7361-0b4b-48a2-bc61-8085b2dbbcb6 req-cb5c2fe1-e063-4db6-8717-73af1c5e6adc service nova] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Received event network-vif-plugged-eba45f53-1f29-46b1-8285-33b5aacbcc43 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 880.543165] env[65522]: DEBUG oslo_concurrency.lockutils [req-792b7361-0b4b-48a2-bc61-8085b2dbbcb6 req-cb5c2fe1-e063-4db6-8717-73af1c5e6adc service nova] Acquiring lock "308330f8-1219-4c02-a129-22b29f0e33b2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 880.543165] env[65522]: DEBUG oslo_concurrency.lockutils [req-792b7361-0b4b-48a2-bc61-8085b2dbbcb6 req-cb5c2fe1-e063-4db6-8717-73af1c5e6adc service nova] Lock "308330f8-1219-4c02-a129-22b29f0e33b2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 880.543165] env[65522]: DEBUG oslo_concurrency.lockutils [req-792b7361-0b4b-48a2-bc61-8085b2dbbcb6 req-cb5c2fe1-e063-4db6-8717-73af1c5e6adc service nova] Lock "308330f8-1219-4c02-a129-22b29f0e33b2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.543165] env[65522]: DEBUG nova.compute.manager [req-792b7361-0b4b-48a2-bc61-8085b2dbbcb6 req-cb5c2fe1-e063-4db6-8717-73af1c5e6adc service nova] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] No waiting events found dispatching network-vif-plugged-eba45f53-1f29-46b1-8285-33b5aacbcc43 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 880.543165] env[65522]: WARNING nova.compute.manager [req-792b7361-0b4b-48a2-bc61-8085b2dbbcb6 req-cb5c2fe1-e063-4db6-8717-73af1c5e6adc service nova] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Received unexpected event network-vif-plugged-eba45f53-1f29-46b1-8285-33b5aacbcc43 for instance with vm_state building and task_state spawning. [ 880.648060] env[65522]: DEBUG oslo_vmware.api [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114208, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.738723] env[65522]: DEBUG nova.network.neutron [-] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 880.778812] env[65522]: DEBUG nova.compute.utils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 880.786363] env[65522]: DEBUG nova.compute.manager [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 880.786363] env[65522]: DEBUG nova.network.neutron [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 880.787346] env[65522]: WARNING neutronclient.v2_0.client [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 880.787346] env[65522]: WARNING neutronclient.v2_0.client [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 880.788044] env[65522]: WARNING openstack [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 880.788763] env[65522]: WARNING openstack [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 880.798897] env[65522]: WARNING neutronclient.v2_0.client [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 880.799481] env[65522]: WARNING openstack [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 880.799823] env[65522]: WARNING openstack [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 880.829651] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31951f0d-04b1-4265-8d07-bd58f6fcddb5 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "1e7b8237-34ea-479e-b5a7-b1846661d61d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.250s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.878741] env[65522]: DEBUG nova.policy [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f51006be1b2e401a8dd5504e0e481810', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ea0fab7de6c47eeaf3bdde413ae3901', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 880.941967] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "refresh_cache-308330f8-1219-4c02-a129-22b29f0e33b2" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.942602] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquired lock "refresh_cache-308330f8-1219-4c02-a129-22b29f0e33b2" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 880.942602] env[65522]: DEBUG nova.network.neutron [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 881.086505] env[65522]: WARNING openstack [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.087543] env[65522]: WARNING openstack [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.102232] env[65522]: DEBUG nova.network.neutron [-] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 881.148870] env[65522]: DEBUG oslo_vmware.api [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114208, 'name': ReconfigVM_Task, 'duration_secs': 0.873915} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.151804] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Reconfigured VM instance instance-0000002e to attach disk [datastore2] volume-33425e73-8f97-4928-afed-0733d11034fb/volume-33425e73-8f97-4928-afed-0733d11034fb.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.157849] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a391aa4c-7569-42f9-ae29-5ff9dc4fb5ae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.178056] env[65522]: DEBUG oslo_vmware.api [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 881.178056] env[65522]: value = "task-5114209" [ 881.178056] env[65522]: _type = "Task" [ 881.178056] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.189946] env[65522]: DEBUG oslo_vmware.api [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114209, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.191252] env[65522]: WARNING neutronclient.v2_0.client [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 881.191907] env[65522]: WARNING openstack [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.192356] env[65522]: WARNING openstack [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.243882] env[65522]: INFO nova.compute.manager [-] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Took 1.38 seconds to deallocate network for instance. [ 881.288309] env[65522]: DEBUG nova.compute.manager [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 881.292443] env[65522]: DEBUG nova.network.neutron [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Successfully created port: a5bdb6f4-7094-4bee-9867-c1d2e683d2f3 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 881.316364] env[65522]: DEBUG nova.network.neutron [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updated VIF entry in instance network info cache for port c6221fb0-8c5b-478e-937e-3997556d0aaa. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 881.316811] env[65522]: DEBUG nova.network.neutron [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updating instance_info_cache with network_info: [{"id": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "address": "fa:16:3e:8e:34:0c", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6221fb0-8c", "ovs_interfaceid": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 881.354648] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64affd0c-3fa2-4fa0-bb8f-4cb4af807f46 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.363606] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5166403a-2f26-4120-bc7f-188081f38b26 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.400833] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4059ae8e-d1bd-464a-8e68-5db7bcba1f64 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.406705] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] Acquiring lock "refresh_cache-1e7b8237-34ea-479e-b5a7-b1846661d61d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.406970] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] Acquired lock "refresh_cache-1e7b8237-34ea-479e-b5a7-b1846661d61d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 881.407052] env[65522]: DEBUG nova.network.neutron [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 881.412944] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335cd118-2de8-48ca-9e86-db54d7ada8dd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.430366] env[65522]: DEBUG nova.compute.provider_tree [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.448438] env[65522]: WARNING openstack [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.449329] env[65522]: WARNING openstack [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.485435] env[65522]: DEBUG nova.network.neutron [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 881.515345] env[65522]: WARNING openstack [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.515712] env[65522]: WARNING openstack [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.594549] env[65522]: WARNING neutronclient.v2_0.client [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 881.595194] env[65522]: WARNING openstack [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.596116] env[65522]: WARNING openstack [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.604839] env[65522]: INFO nova.compute.manager [-] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Took 1.26 seconds to deallocate network for instance. [ 881.688098] env[65522]: DEBUG oslo_vmware.api [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114209, 'name': ReconfigVM_Task, 'duration_secs': 0.258698} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.688464] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994827', 'volume_id': '33425e73-8f97-4928-afed-0733d11034fb', 'name': 'volume-33425e73-8f97-4928-afed-0733d11034fb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1eebb320-30e2-4d8b-b1fd-19e31fb35b77', 'attached_at': '', 'detached_at': '', 'volume_id': '33425e73-8f97-4928-afed-0733d11034fb', 'serial': '33425e73-8f97-4928-afed-0733d11034fb'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 881.752851] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 881.820261] env[65522]: DEBUG oslo_concurrency.lockutils [req-653ec5b8-381c-452c-a232-3f3944a788b4 req-9e93a642-bae1-4dca-8cff-ead9ce441cbb service nova] Releasing lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 881.868605] env[65522]: DEBUG nova.network.neutron [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Updating instance_info_cache with network_info: [{"id": "eba45f53-1f29-46b1-8285-33b5aacbcc43", "address": "fa:16:3e:f5:ef:15", "network": {"id": "472285e2-f5f3-4616-92f7-4c0144d833d4", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-945513036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2bd38f300424db78568e0b8d9982f27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeba45f53-1f", "ovs_interfaceid": "eba45f53-1f29-46b1-8285-33b5aacbcc43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 881.909794] env[65522]: WARNING neutronclient.v2_0.client [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 881.910598] env[65522]: WARNING openstack [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 881.910991] env[65522]: WARNING openstack [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 881.935457] env[65522]: DEBUG nova.scheduler.client.report [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 882.112338] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 882.120129] env[65522]: WARNING openstack [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 882.120722] env[65522]: WARNING openstack [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 882.246169] env[65522]: WARNING neutronclient.v2_0.client [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.246740] env[65522]: WARNING openstack [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 882.247170] env[65522]: WARNING openstack [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 882.257759] env[65522]: DEBUG nova.objects.instance [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lazy-loading 'flavor' on Instance uuid 1eebb320-30e2-4d8b-b1fd-19e31fb35b77 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.302131] env[65522]: DEBUG nova.compute.manager [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 882.330887] env[65522]: DEBUG nova.virt.hardware [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 882.331203] env[65522]: DEBUG nova.virt.hardware [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 882.331360] env[65522]: DEBUG nova.virt.hardware [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 882.331534] env[65522]: DEBUG nova.virt.hardware [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 882.331678] env[65522]: DEBUG nova.virt.hardware [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 882.331822] env[65522]: DEBUG nova.virt.hardware [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 882.332082] env[65522]: DEBUG nova.virt.hardware [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 882.332220] env[65522]: DEBUG nova.virt.hardware [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 882.332400] env[65522]: DEBUG nova.virt.hardware [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 882.332563] env[65522]: DEBUG nova.virt.hardware [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 882.332733] env[65522]: DEBUG nova.virt.hardware [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 882.333931] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b799b16c-6600-4fd4-91a4-51c1ab07f215 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.345029] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1848928d-0c34-43bc-b61f-ad825f3f59b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.372811] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Releasing lock "refresh_cache-308330f8-1219-4c02-a129-22b29f0e33b2" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 882.373275] env[65522]: DEBUG nova.compute.manager [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Instance network_info: |[{"id": "eba45f53-1f29-46b1-8285-33b5aacbcc43", "address": "fa:16:3e:f5:ef:15", "network": {"id": "472285e2-f5f3-4616-92f7-4c0144d833d4", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-945513036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2bd38f300424db78568e0b8d9982f27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeba45f53-1f", "ovs_interfaceid": "eba45f53-1f29-46b1-8285-33b5aacbcc43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 882.373798] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f5:ef:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eba45f53-1f29-46b1-8285-33b5aacbcc43', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 882.382115] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 882.382412] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 882.382649] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ab90e153-cd99-4334-b632-416672f51919 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.401198] env[65522]: DEBUG nova.network.neutron [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Updating instance_info_cache with network_info: [{"id": "0750570e-fa7b-4296-a6f5-dec88fcbbab6", "address": "fa:16:3e:40:ce:37", "network": {"id": "dc9d8462-3539-4cc5-b21a-e1136a8df8e6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1580876213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b5c14a856d454048856539895f220c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0750570e-fa", "ovs_interfaceid": "0750570e-fa7b-4296-a6f5-dec88fcbbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 882.410651] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 882.410651] env[65522]: value = "task-5114210" [ 882.410651] env[65522]: _type = "Task" [ 882.410651] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.422215] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114210, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.439908] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.170s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 882.440572] env[65522]: DEBUG nova.compute.manager [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 882.443402] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 16.158s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 882.514947] env[65522]: DEBUG nova.compute.manager [req-75fc1a00-0c1d-4892-801b-9e0c2ac9d01f req-32fc8f93-3f5c-46b7-9464-3b0c70837858 service nova] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Received event network-vif-deleted-dd99a89e-fb55-4af2-94d6-979e1b1cae45 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 882.515230] env[65522]: DEBUG nova.compute.manager [req-75fc1a00-0c1d-4892-801b-9e0c2ac9d01f req-32fc8f93-3f5c-46b7-9464-3b0c70837858 service nova] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Received event network-vif-deleted-e296c3b2-a6e6-4f3a-8fc8-9a208bb85470 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 882.667963] env[65522]: DEBUG nova.compute.manager [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Received event network-changed-eba45f53-1f29-46b1-8285-33b5aacbcc43 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 882.668289] env[65522]: DEBUG nova.compute.manager [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Refreshing instance network info cache due to event network-changed-eba45f53-1f29-46b1-8285-33b5aacbcc43. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 882.668625] env[65522]: DEBUG oslo_concurrency.lockutils [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] Acquiring lock "refresh_cache-308330f8-1219-4c02-a129-22b29f0e33b2" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.669095] env[65522]: DEBUG oslo_concurrency.lockutils [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] Acquired lock "refresh_cache-308330f8-1219-4c02-a129-22b29f0e33b2" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 882.669380] env[65522]: DEBUG nova.network.neutron [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Refreshing network info cache for port eba45f53-1f29-46b1-8285-33b5aacbcc43 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 882.764040] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57176b65-bf65-41cb-8526-a7338cbdc864 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.353s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 882.877203] env[65522]: DEBUG nova.network.neutron [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Successfully updated port: a5bdb6f4-7094-4bee-9867-c1d2e683d2f3 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 882.903713] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] Releasing lock "refresh_cache-1e7b8237-34ea-479e-b5a7-b1846661d61d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 882.905287] env[65522]: DEBUG nova.compute.manager [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Inject network info {{(pid=65522) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7794}} [ 882.905287] env[65522]: DEBUG nova.compute.manager [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] network_info to inject: |[{"id": "0750570e-fa7b-4296-a6f5-dec88fcbbab6", "address": "fa:16:3e:40:ce:37", "network": {"id": "dc9d8462-3539-4cc5-b21a-e1136a8df8e6", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1580876213-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b5c14a856d454048856539895f220c2c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0750570e-fa", "ovs_interfaceid": "0750570e-fa7b-4296-a6f5-dec88fcbbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7795}} [ 882.910598] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Reconfiguring VM instance to set the machine id {{(pid=65522) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 882.911409] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92de2be2-de7f-46d3-962c-b3cf7388a543 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.936307] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114210, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.937873] env[65522]: DEBUG oslo_vmware.api [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] Waiting for the task: (returnval){ [ 882.937873] env[65522]: value = "task-5114211" [ 882.937873] env[65522]: _type = "Task" [ 882.937873] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.950029] env[65522]: INFO nova.compute.claims [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.955021] env[65522]: DEBUG nova.compute.utils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 882.956775] env[65522]: DEBUG oslo_vmware.api [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] Task: {'id': task-5114211, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.957642] env[65522]: DEBUG nova.compute.manager [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 882.957784] env[65522]: DEBUG nova.network.neutron [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 882.958223] env[65522]: WARNING neutronclient.v2_0.client [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.958596] env[65522]: WARNING neutronclient.v2_0.client [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.959526] env[65522]: WARNING openstack [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 882.959841] env[65522]: WARNING openstack [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 883.019380] env[65522]: DEBUG nova.policy [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e39a3da2fd584e6ca40574518881c3fb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cb18797424f3489ab6747cb859d7ff23', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 883.173595] env[65522]: WARNING neutronclient.v2_0.client [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 883.174744] env[65522]: WARNING openstack [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 883.174835] env[65522]: WARNING openstack [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 883.296424] env[65522]: WARNING openstack [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 883.296824] env[65522]: WARNING openstack [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 883.355022] env[65522]: DEBUG nova.network.neutron [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Successfully created port: 237a8db4-69b6-49b1-8828-e7aa07977495 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 883.380751] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "refresh_cache-4264998b-949d-4b01-8832-fb1df6214403" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.380751] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "refresh_cache-4264998b-949d-4b01-8832-fb1df6214403" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 883.380751] env[65522]: DEBUG nova.network.neutron [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 883.407996] env[65522]: WARNING neutronclient.v2_0.client [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 883.408718] env[65522]: WARNING openstack [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 883.409108] env[65522]: WARNING openstack [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 883.440398] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114210, 'name': CreateVM_Task, 'duration_secs': 0.528594} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.443944] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 883.444572] env[65522]: WARNING neutronclient.v2_0.client [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 883.444951] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.445125] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 883.446936] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 883.446936] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4fe91ea2-c97b-4ba5-a621-10ec8b218b20 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.451812] env[65522]: DEBUG oslo_vmware.api [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] Task: {'id': task-5114211, 'name': ReconfigVM_Task, 'duration_secs': 0.208438} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.452213] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f7f8a5fb-31a7-4ac5-9b0b-9589628480f5 tempest-ServersAdminTestJSON-760417853 tempest-ServersAdminTestJSON-760417853-project-admin] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Reconfigured VM instance to set the machine id {{(pid=65522) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 883.455405] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 883.455405] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520ede6e-97b2-65cc-82d7-cd79037c6932" [ 883.455405] env[65522]: _type = "Task" [ 883.455405] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.460764] env[65522]: INFO nova.compute.resource_tracker [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating resource usage from migration ed83be76-1d3a-4a25-85ce-9cf749f05abc [ 883.463475] env[65522]: DEBUG nova.compute.manager [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 883.474580] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520ede6e-97b2-65cc-82d7-cd79037c6932, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.529063] env[65522]: DEBUG nova.network.neutron [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Updated VIF entry in instance network info cache for port eba45f53-1f29-46b1-8285-33b5aacbcc43. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 883.529458] env[65522]: DEBUG nova.network.neutron [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Updating instance_info_cache with network_info: [{"id": "eba45f53-1f29-46b1-8285-33b5aacbcc43", "address": "fa:16:3e:f5:ef:15", "network": {"id": "472285e2-f5f3-4616-92f7-4c0144d833d4", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-945513036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2bd38f300424db78568e0b8d9982f27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeba45f53-1f", "ovs_interfaceid": "eba45f53-1f29-46b1-8285-33b5aacbcc43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 883.886804] env[65522]: WARNING openstack [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 883.887496] env[65522]: WARNING openstack [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 883.956977] env[65522]: DEBUG nova.network.neutron [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 883.974459] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520ede6e-97b2-65cc-82d7-cd79037c6932, 'name': SearchDatastore_Task, 'duration_secs': 0.026004} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.977562] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 883.977797] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.978075] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.978227] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 883.978405] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.978846] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-14131bf9-6b4d-4062-b002-8165a691b06d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.988607] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.988736] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 883.992293] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09b89cc6-0756-4f96-ad05-8b10c487a2f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.998720] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 883.998720] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52176f4e-85be-426d-2f31-261faced0f86" [ 883.998720] env[65522]: _type = "Task" [ 883.998720] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.008505] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52176f4e-85be-426d-2f31-261faced0f86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.039066] env[65522]: DEBUG oslo_concurrency.lockutils [req-97e22b6c-42ee-4eb9-a89f-daa80506e640 req-f35b35e6-83eb-4572-9061-cb91328f71fc service nova] Releasing lock "refresh_cache-308330f8-1219-4c02-a129-22b29f0e33b2" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 884.043672] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d54aef-7626-4ff3-aecc-4df04251b217 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.047826] env[65522]: WARNING openstack [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 884.048235] env[65522]: WARNING openstack [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 884.061036] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e6a4ce-459c-46c8-b2a9-5d09c7010017 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.098031] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-656d733f-9d5e-4c27-8736-c46028508c0c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.107018] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-041ed14b-d042-4b76-b61d-6791503e9576 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.128992] env[65522]: DEBUG nova.compute.provider_tree [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.146670] env[65522]: WARNING neutronclient.v2_0.client [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 884.147543] env[65522]: WARNING openstack [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 884.147926] env[65522]: WARNING openstack [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 884.274784] env[65522]: INFO nova.compute.manager [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Rescuing [ 884.275101] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.275262] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 884.275430] env[65522]: DEBUG nova.network.neutron [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 884.317470] env[65522]: DEBUG nova.network.neutron [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Updating instance_info_cache with network_info: [{"id": "a5bdb6f4-7094-4bee-9867-c1d2e683d2f3", "address": "fa:16:3e:22:15:17", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5bdb6f4-70", "ovs_interfaceid": "a5bdb6f4-7094-4bee-9867-c1d2e683d2f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 884.479138] env[65522]: DEBUG nova.compute.manager [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 884.502630] env[65522]: DEBUG nova.virt.hardware [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 884.502971] env[65522]: DEBUG nova.virt.hardware [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 884.503074] env[65522]: DEBUG nova.virt.hardware [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 884.503286] env[65522]: DEBUG nova.virt.hardware [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 884.503430] env[65522]: DEBUG nova.virt.hardware [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 884.503576] env[65522]: DEBUG nova.virt.hardware [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 884.503784] env[65522]: DEBUG nova.virt.hardware [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.503939] env[65522]: DEBUG nova.virt.hardware [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 884.504127] env[65522]: DEBUG nova.virt.hardware [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 884.504293] env[65522]: DEBUG nova.virt.hardware [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 884.504460] env[65522]: DEBUG nova.virt.hardware [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 884.508497] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8978ad-4936-451f-8af8-5abd2eaa471f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.519986] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52176f4e-85be-426d-2f31-261faced0f86, 'name': SearchDatastore_Task, 'duration_secs': 0.012132} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.522384] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27ab3fc-73b5-4493-a7e6-202bbb5ffd1d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.526750] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a48bcf7-a792-4981-b9bd-4e5441589fd9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.541407] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 884.541407] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523afd98-1c1d-47cf-ab98-37718821f5ae" [ 884.541407] env[65522]: _type = "Task" [ 884.541407] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.552387] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523afd98-1c1d-47cf-ab98-37718821f5ae, 'name': SearchDatastore_Task, 'duration_secs': 0.012675} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.552701] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 884.552807] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 308330f8-1219-4c02-a129-22b29f0e33b2/308330f8-1219-4c02-a129-22b29f0e33b2.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 884.553118] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a9e638dd-ca50-4a31-a74b-d2b37891cc04 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.561709] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 884.561709] env[65522]: value = "task-5114212" [ 884.561709] env[65522]: _type = "Task" [ 884.561709] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.571455] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114212, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.634398] env[65522]: DEBUG nova.scheduler.client.report [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 884.779745] env[65522]: WARNING neutronclient.v2_0.client [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 884.780571] env[65522]: WARNING openstack [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 884.780937] env[65522]: WARNING openstack [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 884.821763] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "refresh_cache-4264998b-949d-4b01-8832-fb1df6214403" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 884.822226] env[65522]: DEBUG nova.compute.manager [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Instance network_info: |[{"id": "a5bdb6f4-7094-4bee-9867-c1d2e683d2f3", "address": "fa:16:3e:22:15:17", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5bdb6f4-70", "ovs_interfaceid": "a5bdb6f4-7094-4bee-9867-c1d2e683d2f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 884.822750] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:15:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5e1049e8-c06b-4c93-a9e1-2cbb530f3f95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a5bdb6f4-7094-4bee-9867-c1d2e683d2f3', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.831179] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Creating folder: Project (3ea0fab7de6c47eeaf3bdde413ae3901). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 884.831508] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a1364f19-e978-40bc-84d6-e81e5cf63012 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.845862] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Created folder: Project (3ea0fab7de6c47eeaf3bdde413ae3901) in parent group-v994660. [ 884.846302] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Creating folder: Instances. Parent ref: group-v994829. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 884.846605] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7db1419b-2a04-4781-925a-c163a3077da9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.860313] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Created folder: Instances in parent group-v994829. [ 884.860772] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 884.861031] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 884.861363] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48810ceb-8054-49ed-b91a-27f39da4a742 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.886450] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.886450] env[65522]: value = "task-5114215" [ 884.886450] env[65522]: _type = "Task" [ 884.886450] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.896797] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114215, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.956411] env[65522]: INFO nova.compute.manager [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Rebuilding instance [ 885.013658] env[65522]: DEBUG nova.compute.manager [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 885.014784] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68dd940c-4307-4ee3-a7d9-cb25c1c6b3e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.019610] env[65522]: DEBUG nova.network.neutron [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Successfully updated port: 237a8db4-69b6-49b1-8828-e7aa07977495 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 885.053375] env[65522]: DEBUG nova.compute.manager [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Received event network-vif-plugged-a5bdb6f4-7094-4bee-9867-c1d2e683d2f3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 885.053661] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] Acquiring lock "4264998b-949d-4b01-8832-fb1df6214403-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.053889] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] Lock "4264998b-949d-4b01-8832-fb1df6214403-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 885.054094] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] Lock "4264998b-949d-4b01-8832-fb1df6214403-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 885.054273] env[65522]: DEBUG nova.compute.manager [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] [instance: 4264998b-949d-4b01-8832-fb1df6214403] No waiting events found dispatching network-vif-plugged-a5bdb6f4-7094-4bee-9867-c1d2e683d2f3 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 885.054455] env[65522]: WARNING nova.compute.manager [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Received unexpected event network-vif-plugged-a5bdb6f4-7094-4bee-9867-c1d2e683d2f3 for instance with vm_state building and task_state spawning. [ 885.054589] env[65522]: DEBUG nova.compute.manager [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Received event network-changed-a5bdb6f4-7094-4bee-9867-c1d2e683d2f3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 885.054816] env[65522]: DEBUG nova.compute.manager [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Refreshing instance network info cache due to event network-changed-a5bdb6f4-7094-4bee-9867-c1d2e683d2f3. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 885.055033] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] Acquiring lock "refresh_cache-4264998b-949d-4b01-8832-fb1df6214403" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.055180] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] Acquired lock "refresh_cache-4264998b-949d-4b01-8832-fb1df6214403" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.055357] env[65522]: DEBUG nova.network.neutron [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Refreshing network info cache for port a5bdb6f4-7094-4bee-9867-c1d2e683d2f3 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 885.078600] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114212, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.140234] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.697s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 885.140510] env[65522]: INFO nova.compute.manager [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Migrating [ 885.151879] env[65522]: WARNING openstack [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.153219] env[65522]: WARNING openstack [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.160559] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 18.160s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 885.235692] env[65522]: WARNING neutronclient.v2_0.client [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 885.235692] env[65522]: WARNING openstack [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.235993] env[65522]: WARNING openstack [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.378096] env[65522]: DEBUG nova.network.neutron [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Updating instance_info_cache with network_info: [{"id": "e12c2307-4963-4a5a-af0f-19369217f430", "address": "fa:16:3e:9f:b8:7d", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape12c2307-49", "ovs_interfaceid": "e12c2307-4963-4a5a-af0f-19369217f430", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 885.402299] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114215, 'name': CreateVM_Task, 'duration_secs': 0.435091} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.403585] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.404117] env[65522]: WARNING neutronclient.v2_0.client [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 885.404612] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.405344] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.405344] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 885.405599] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5ea8610-dc99-4036-971c-ed120624c309 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.411801] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 885.411801] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ffe836-0b9e-8c1a-e0a8-3a0f3c41a07d" [ 885.411801] env[65522]: _type = "Task" [ 885.411801] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.421663] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ffe836-0b9e-8c1a-e0a8-3a0f3c41a07d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.524200] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Acquiring lock "refresh_cache-9da312a0-c6eb-4903-a7d8-e178aa54d2ea" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.524200] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Acquired lock "refresh_cache-9da312a0-c6eb-4903-a7d8-e178aa54d2ea" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.524200] env[65522]: DEBUG nova.network.neutron [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 885.568181] env[65522]: WARNING neutronclient.v2_0.client [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 885.568669] env[65522]: WARNING openstack [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.569042] env[65522]: WARNING openstack [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.583331] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114212, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.671933] env[65522]: WARNING openstack [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.672465] env[65522]: WARNING openstack [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.678961] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.679142] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.679316] env[65522]: DEBUG nova.network.neutron [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 885.738573] env[65522]: WARNING neutronclient.v2_0.client [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 885.739270] env[65522]: WARNING openstack [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.739671] env[65522]: WARNING openstack [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.840351] env[65522]: DEBUG nova.network.neutron [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Updated VIF entry in instance network info cache for port a5bdb6f4-7094-4bee-9867-c1d2e683d2f3. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 885.840621] env[65522]: DEBUG nova.network.neutron [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Updating instance_info_cache with network_info: [{"id": "a5bdb6f4-7094-4bee-9867-c1d2e683d2f3", "address": "fa:16:3e:22:15:17", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5bdb6f4-70", "ovs_interfaceid": "a5bdb6f4-7094-4bee-9867-c1d2e683d2f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 885.883426] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 885.931901] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ffe836-0b9e-8c1a-e0a8-3a0f3c41a07d, 'name': SearchDatastore_Task, 'duration_secs': 0.011741} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.932396] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 885.932775] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.933186] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.933521] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.933762] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.934615] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8707df0b-6572-4620-ad40-f044b9d39830 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.952563] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.953087] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 885.957470] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2eb72856-d3e9-4160-a782-77189bf77943 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.965736] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 885.965736] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5253160c-9e60-c0cc-cbfa-abb2509d060b" [ 885.965736] env[65522]: _type = "Task" [ 885.965736] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.980944] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5253160c-9e60-c0cc-cbfa-abb2509d060b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.028968] env[65522]: WARNING openstack [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.029684] env[65522]: WARNING openstack [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.039195] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 886.039195] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e09d4a69-9261-4d9a-904f-89c6ed0b2e76 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.048488] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 886.048488] env[65522]: value = "task-5114216" [ 886.048488] env[65522]: _type = "Task" [ 886.048488] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.065263] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114216, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.080496] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114212, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.081585] env[65522]: DEBUG nova.network.neutron [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 886.123613] env[65522]: WARNING openstack [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.124246] env[65522]: WARNING openstack [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.184018] env[65522]: WARNING neutronclient.v2_0.client [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 886.184736] env[65522]: WARNING openstack [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.185134] env[65522]: WARNING openstack [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.194365] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a66282-8671-4a96-b065-4adfe5e653e3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.203686] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9217bdd2-41ae-4b7d-b9d8-7bbe29cc8692 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.261447] env[65522]: WARNING neutronclient.v2_0.client [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 886.261825] env[65522]: WARNING openstack [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.262197] env[65522]: WARNING openstack [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.273266] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c812dbb-057c-45ef-9b43-4adbeda01430 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.282097] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce53a21-94f3-41ae-b91d-a61abb246e64 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.298421] env[65522]: DEBUG nova.compute.provider_tree [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.345359] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f6185ae-877c-4bf0-b04d-6502ac201156 req-fda9c369-7026-40cf-9e1f-a64e25c58d35 service nova] Releasing lock "refresh_cache-4264998b-949d-4b01-8832-fb1df6214403" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 886.453081] env[65522]: DEBUG nova.network.neutron [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Updating instance_info_cache with network_info: [{"id": "237a8db4-69b6-49b1-8828-e7aa07977495", "address": "fa:16:3e:ed:01:5b", "network": {"id": "f9c988f4-de84-44bf-8bb1-809fc764b925", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1490980998-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cb18797424f3489ab6747cb859d7ff23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cbd5e0e-9116-46f1-9748-13a73d2d7e75", "external-id": "nsx-vlan-transportzone-690", "segmentation_id": 690, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap237a8db4-69", "ovs_interfaceid": "237a8db4-69b6-49b1-8828-e7aa07977495", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 886.480786] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5253160c-9e60-c0cc-cbfa-abb2509d060b, 'name': SearchDatastore_Task, 'duration_secs': 0.019204} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.480786] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7eaa3a40-fcad-419f-a1ca-7a25fa0e5612 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.486345] env[65522]: WARNING openstack [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.486876] env[65522]: WARNING openstack [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.496437] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 886.496437] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52caab0e-4673-1c11-7916-e797f9434fb3" [ 886.496437] env[65522]: _type = "Task" [ 886.496437] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.510759] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52caab0e-4673-1c11-7916-e797f9434fb3, 'name': SearchDatastore_Task, 'duration_secs': 0.014547} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.510759] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 886.510759] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 4264998b-949d-4b01-8832-fb1df6214403/4264998b-949d-4b01-8832-fb1df6214403.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 886.510759] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c053bdaa-5131-4400-addd-f17699337b9a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.519576] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 886.519576] env[65522]: value = "task-5114217" [ 886.519576] env[65522]: _type = "Task" [ 886.519576] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.536529] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114217, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.561555] env[65522]: WARNING neutronclient.v2_0.client [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 886.562331] env[65522]: WARNING openstack [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 886.562796] env[65522]: WARNING openstack [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.584255] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114216, 'name': PowerOffVM_Task, 'duration_secs': 0.456132} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.584970] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 886.584970] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 886.587112] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a46c26d4-1887-498d-9b93-89c9a845bc2b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.593104] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114212, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.679688} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.593838] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 308330f8-1219-4c02-a129-22b29f0e33b2/308330f8-1219-4c02-a129-22b29f0e33b2.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 886.594069] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 886.594369] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-722e60ec-5fdb-4810-9780-e22c901b05a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.601738] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 886.602644] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d073a2bb-fa5a-4a48-b192-d46a89ba48c8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.609543] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 886.609543] env[65522]: value = "task-5114218" [ 886.609543] env[65522]: _type = "Task" [ 886.609543] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.621563] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114218, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.691092] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 886.691092] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 886.691092] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleting the datastore file [datastore2] bb20d99f-fc49-47bf-a6b7-74ea966b71eb {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 886.691561] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-564ea4ae-9efe-48e1-84ba-f22dbb7e9db3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.702924] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 886.702924] env[65522]: value = "task-5114220" [ 886.702924] env[65522]: _type = "Task" [ 886.702924] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.714852] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114220, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.717734] env[65522]: DEBUG nova.network.neutron [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance_info_cache with network_info: [{"id": "3f79042c-23dd-4fc0-b3cf-1e2884edabc1", "address": "fa:16:3e:f7:f5:9e", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.131", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f79042c-23", "ovs_interfaceid": "3f79042c-23dd-4fc0-b3cf-1e2884edabc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 886.808028] env[65522]: DEBUG nova.scheduler.client.report [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 886.961411] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Releasing lock "refresh_cache-9da312a0-c6eb-4903-a7d8-e178aa54d2ea" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 886.961411] env[65522]: DEBUG nova.compute.manager [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Instance network_info: |[{"id": "237a8db4-69b6-49b1-8828-e7aa07977495", "address": "fa:16:3e:ed:01:5b", "network": {"id": "f9c988f4-de84-44bf-8bb1-809fc764b925", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1490980998-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cb18797424f3489ab6747cb859d7ff23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cbd5e0e-9116-46f1-9748-13a73d2d7e75", "external-id": "nsx-vlan-transportzone-690", "segmentation_id": 690, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap237a8db4-69", "ovs_interfaceid": "237a8db4-69b6-49b1-8828-e7aa07977495", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 886.961411] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ed:01:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1cbd5e0e-9116-46f1-9748-13a73d2d7e75', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '237a8db4-69b6-49b1-8828-e7aa07977495', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 886.967831] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Creating folder: Project (cb18797424f3489ab6747cb859d7ff23). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 886.968745] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c0988e17-6131-4f55-b057-a642cee6cf8e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.984023] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Created folder: Project (cb18797424f3489ab6747cb859d7ff23) in parent group-v994660. [ 886.984023] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Creating folder: Instances. Parent ref: group-v994832. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 886.984023] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6e11ff48-4b8d-4a80-829f-05a406735749 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.994966] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Created folder: Instances in parent group-v994832. [ 886.995445] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 886.996032] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 886.996438] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23e5da44-94ab-4d10-9cdb-57c92bd1e900 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.020543] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.020543] env[65522]: value = "task-5114223" [ 887.020543] env[65522]: _type = "Task" [ 887.020543] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.038033] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114223, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.038869] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114217, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.125833] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114218, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0915} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.126462] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 887.127969] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4a1bb3-543d-4656-8675-121382ff6a89 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.162796] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 308330f8-1219-4c02-a129-22b29f0e33b2/308330f8-1219-4c02-a129-22b29f0e33b2.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 887.163529] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ddbdddd7-55c5-4340-a3d7-ac6c91407aba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.188033] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Acquiring lock "18568baa-83de-4b54-addf-dee66211c8d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 887.188033] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Lock "18568baa-83de-4b54-addf-dee66211c8d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 887.188033] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Acquiring lock "18568baa-83de-4b54-addf-dee66211c8d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 887.188033] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Lock "18568baa-83de-4b54-addf-dee66211c8d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 887.188033] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Lock "18568baa-83de-4b54-addf-dee66211c8d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 887.191867] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 887.191867] env[65522]: value = "task-5114224" [ 887.191867] env[65522]: _type = "Task" [ 887.191867] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.192890] env[65522]: INFO nova.compute.manager [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Terminating instance [ 887.212960] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114224, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.223667] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Releasing lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 887.228966] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114220, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152606} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.228966] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 887.228966] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 887.228966] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 887.421427] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.421800] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-979d49cf-ea06-468e-8c98-5898961d54f8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.428666] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 887.428666] env[65522]: value = "task-5114225" [ 887.428666] env[65522]: _type = "Task" [ 887.428666] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.440779] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114225, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.495888] env[65522]: DEBUG nova.compute.manager [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Received event network-vif-plugged-237a8db4-69b6-49b1-8828-e7aa07977495 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 887.495888] env[65522]: DEBUG oslo_concurrency.lockutils [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] Acquiring lock "9da312a0-c6eb-4903-a7d8-e178aa54d2ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 887.495996] env[65522]: DEBUG oslo_concurrency.lockutils [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] Lock "9da312a0-c6eb-4903-a7d8-e178aa54d2ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 887.496223] env[65522]: DEBUG oslo_concurrency.lockutils [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] Lock "9da312a0-c6eb-4903-a7d8-e178aa54d2ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 887.496443] env[65522]: DEBUG nova.compute.manager [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] No waiting events found dispatching network-vif-plugged-237a8db4-69b6-49b1-8828-e7aa07977495 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 887.496692] env[65522]: WARNING nova.compute.manager [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Received unexpected event network-vif-plugged-237a8db4-69b6-49b1-8828-e7aa07977495 for instance with vm_state building and task_state spawning. [ 887.496912] env[65522]: DEBUG nova.compute.manager [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Received event network-changed-237a8db4-69b6-49b1-8828-e7aa07977495 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 887.497112] env[65522]: DEBUG nova.compute.manager [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Refreshing instance network info cache due to event network-changed-237a8db4-69b6-49b1-8828-e7aa07977495. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 887.497349] env[65522]: DEBUG oslo_concurrency.lockutils [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] Acquiring lock "refresh_cache-9da312a0-c6eb-4903-a7d8-e178aa54d2ea" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.497618] env[65522]: DEBUG oslo_concurrency.lockutils [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] Acquired lock "refresh_cache-9da312a0-c6eb-4903-a7d8-e178aa54d2ea" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 887.497817] env[65522]: DEBUG nova.network.neutron [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Refreshing network info cache for port 237a8db4-69b6-49b1-8828-e7aa07977495 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 887.538299] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114217, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.762672} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.542249] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 4264998b-949d-4b01-8832-fb1df6214403/4264998b-949d-4b01-8832-fb1df6214403.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 887.542566] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 887.543182] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114223, 'name': CreateVM_Task, 'duration_secs': 0.438332} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.543392] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-db730870-82c8-4851-b9b1-6cbed9cfadd7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.545964] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 887.547038] env[65522]: WARNING neutronclient.v2_0.client [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 887.547038] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.547191] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 887.547491] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 887.548345] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2e870cb-ea23-407b-815d-bf7334ae3537 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.553499] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 887.553499] env[65522]: value = "task-5114226" [ 887.553499] env[65522]: _type = "Task" [ 887.553499] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.554893] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Waiting for the task: (returnval){ [ 887.554893] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52325357-87fe-22b9-dffe-c977077d2bf6" [ 887.554893] env[65522]: _type = "Task" [ 887.554893] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.575625] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114226, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.578481] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52325357-87fe-22b9-dffe-c977077d2bf6, 'name': SearchDatastore_Task, 'duration_secs': 0.016242} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.578820] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 887.579027] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 887.579358] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.579500] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 887.579675] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 887.580022] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c174f0c-13ca-400c-bdab-073c5d3c9fff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.590817] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 887.590994] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 887.591871] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c593176d-acd1-48fc-ab2d-5a8b87183230 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.599427] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Waiting for the task: (returnval){ [ 887.599427] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dda8c3-17a5-8252-f7fe-956ef2ddd525" [ 887.599427] env[65522]: _type = "Task" [ 887.599427] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.609484] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dda8c3-17a5-8252-f7fe-956ef2ddd525, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.702080] env[65522]: DEBUG nova.compute.manager [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 887.702414] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 887.706358] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f89b24-deb9-427f-a67e-7d32978b4438 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.709267] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114224, 'name': ReconfigVM_Task, 'duration_secs': 0.387775} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.709422] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 308330f8-1219-4c02-a129-22b29f0e33b2/308330f8-1219-4c02-a129-22b29f0e33b2.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.710482] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b9c3a513-825e-4bc2-9a8b-b39ad22c4506 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.716049] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.717407] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd9cfe39-b832-4e99-96ed-c0365b04ed6f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.719307] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 887.719307] env[65522]: value = "task-5114227" [ 887.719307] env[65522]: _type = "Task" [ 887.719307] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.733688] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114227, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.741632] env[65522]: DEBUG oslo_vmware.api [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Waiting for the task: (returnval){ [ 887.741632] env[65522]: value = "task-5114228" [ 887.741632] env[65522]: _type = "Task" [ 887.741632] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.752813] env[65522]: DEBUG oslo_vmware.api [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114228, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.821946] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.661s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 887.824772] env[65522]: DEBUG oslo_concurrency.lockutils [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.333s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 887.825025] env[65522]: DEBUG nova.objects.instance [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lazy-loading 'resources' on Instance uuid 361e54c7-4309-48b3-ad3e-b2ebc783741c {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.942614] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114225, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.001545] env[65522]: WARNING neutronclient.v2_0.client [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 888.002267] env[65522]: WARNING openstack [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.002618] env[65522]: WARNING openstack [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.064945] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114226, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.133986} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.064945] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.065705] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b45d5c-5f67-456d-b173-6199e2a193e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.088889] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 4264998b-949d-4b01-8832-fb1df6214403/4264998b-949d-4b01-8832-fb1df6214403.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.093181] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c5555e7-9b6e-477a-b9eb-782f392ecfa6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.117993] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dda8c3-17a5-8252-f7fe-956ef2ddd525, 'name': SearchDatastore_Task, 'duration_secs': 0.013628} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.120055] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 888.120055] env[65522]: value = "task-5114229" [ 888.120055] env[65522]: _type = "Task" [ 888.120055] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.120272] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1e79b3f-e383-48dd-b61c-cc59e1ce80f9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.126943] env[65522]: WARNING openstack [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.127431] env[65522]: WARNING openstack [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.140774] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Waiting for the task: (returnval){ [ 888.140774] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e3bd9a-e41d-f5f5-d742-de6e24e4caee" [ 888.140774] env[65522]: _type = "Task" [ 888.140774] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.141445] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114229, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.151234] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e3bd9a-e41d-f5f5-d742-de6e24e4caee, 'name': SearchDatastore_Task, 'duration_secs': 0.012949} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.151352] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 888.151638] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 9da312a0-c6eb-4903-a7d8-e178aa54d2ea/9da312a0-c6eb-4903-a7d8-e178aa54d2ea.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 888.152040] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d409ebe-620d-4cab-adab-764709333bb6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.158213] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Waiting for the task: (returnval){ [ 888.158213] env[65522]: value = "task-5114230" [ 888.158213] env[65522]: _type = "Task" [ 888.158213] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.166506] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114230, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.198946] env[65522]: WARNING neutronclient.v2_0.client [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 888.199667] env[65522]: WARNING openstack [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.200118] env[65522]: WARNING openstack [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.229373] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114227, 'name': Rename_Task, 'duration_secs': 0.163403} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.229555] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 888.229723] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c894af9d-2fce-49ea-9ef6-4ed4b5ac36c5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.238661] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 888.238661] env[65522]: value = "task-5114231" [ 888.238661] env[65522]: _type = "Task" [ 888.238661] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.260334] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114231, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.264141] env[65522]: DEBUG oslo_vmware.api [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114228, 'name': PowerOffVM_Task, 'duration_secs': 0.233757} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.264680] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 888.264935] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 888.265169] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6b93668c-0c79-4dd4-83a6-17b2733fe750 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.277354] env[65522]: DEBUG nova.virt.hardware [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 888.277650] env[65522]: DEBUG nova.virt.hardware [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 888.277810] env[65522]: DEBUG nova.virt.hardware [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 888.278040] env[65522]: DEBUG nova.virt.hardware [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 888.278204] env[65522]: DEBUG nova.virt.hardware [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 888.278352] env[65522]: DEBUG nova.virt.hardware [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 888.278580] env[65522]: DEBUG nova.virt.hardware [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 888.278741] env[65522]: DEBUG nova.virt.hardware [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 888.278905] env[65522]: DEBUG nova.virt.hardware [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 888.279083] env[65522]: DEBUG nova.virt.hardware [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 888.279260] env[65522]: DEBUG nova.virt.hardware [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 888.280224] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2123b8b-98f3-4a62-9796-863831c6abde {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.289194] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07cb9d5f-83d3-4b65-beb0-57f247f2e775 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.296691] env[65522]: DEBUG nova.network.neutron [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Updated VIF entry in instance network info cache for port 237a8db4-69b6-49b1-8828-e7aa07977495. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 888.296691] env[65522]: DEBUG nova.network.neutron [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Updating instance_info_cache with network_info: [{"id": "237a8db4-69b6-49b1-8828-e7aa07977495", "address": "fa:16:3e:ed:01:5b", "network": {"id": "f9c988f4-de84-44bf-8bb1-809fc764b925", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1490980998-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cb18797424f3489ab6747cb859d7ff23", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cbd5e0e-9116-46f1-9748-13a73d2d7e75", "external-id": "nsx-vlan-transportzone-690", "segmentation_id": 690, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap237a8db4-69", "ovs_interfaceid": "237a8db4-69b6-49b1-8828-e7aa07977495", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 888.310501] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:e2:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '652babe2-668c-4364-92c2-8ceaba23c0ac', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.318189] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 888.319179] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 888.319346] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-deddcad6-1d0b-46d0-bd13-01dda393a41a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.346629] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.346629] env[65522]: value = "task-5114233" [ 888.346629] env[65522]: _type = "Task" [ 888.346629] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.355933] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114233, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.400040] env[65522]: INFO nova.scheduler.client.report [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleted allocation for migration 98e33664-f395-480b-a10d-3904a0963458 [ 888.411711] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 888.411845] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 888.412085] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Deleting the datastore file [datastore1] 18568baa-83de-4b54-addf-dee66211c8d7 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 888.412422] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-255414b4-61ad-4264-86ed-2ee92db1ea58 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.430123] env[65522]: DEBUG oslo_vmware.api [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Waiting for the task: (returnval){ [ 888.430123] env[65522]: value = "task-5114234" [ 888.430123] env[65522]: _type = "Task" [ 888.430123] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.449161] env[65522]: DEBUG oslo_vmware.api [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114234, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.457990] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114225, 'name': PowerOffVM_Task, 'duration_secs': 0.52227} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.460426] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 888.464142] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c204a04-6c1a-45c4-9b1a-f7ddf36bd85c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.490655] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9fa3a8-7e4c-49f9-bd54-57a18c64d354 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.529428] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 888.529815] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6db3240-88ad-40d7-8cc4-23ff95920621 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.541416] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 888.541416] env[65522]: value = "task-5114235" [ 888.541416] env[65522]: _type = "Task" [ 888.541416] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.550042] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 888.550536] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.550902] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.551108] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 888.551300] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.554331] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aae47b07-a1d1-4f80-bd43-4f1daeaf6d73 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.565529] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.565529] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 888.566393] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bfeab03-98ac-4e86-b1b1-2756f621dd5f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.576307] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 888.576307] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ab0249-3be1-3834-58a2-115201f3f6ed" [ 888.576307] env[65522]: _type = "Task" [ 888.576307] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.586440] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ab0249-3be1-3834-58a2-115201f3f6ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.636508] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114229, 'name': ReconfigVM_Task, 'duration_secs': 0.338797} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.636919] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 4264998b-949d-4b01-8832-fb1df6214403/4264998b-949d-4b01-8832-fb1df6214403.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.637669] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-314a3f2c-2fc5-4408-beba-6f9a4b4bf146 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.648618] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 888.648618] env[65522]: value = "task-5114236" [ 888.648618] env[65522]: _type = "Task" [ 888.648618] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.662475] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114236, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.677021] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114230, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.751096] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb772c4-53e1-4de0-a279-1c0e3ad0774f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.763064] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114231, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.777442] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance '5e227f6c-9f13-416f-8e6d-2f7d931619fd' progress to 0 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 888.800272] env[65522]: DEBUG oslo_concurrency.lockutils [req-33663a5d-6bbb-4152-8ebf-de6ea8e7ef4f req-efe3f319-ff35-41ed-93cd-d49d3814fcf3 service nova] Releasing lock "refresh_cache-9da312a0-c6eb-4903-a7d8-e178aa54d2ea" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 888.860957] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114233, 'name': CreateVM_Task, 'duration_secs': 0.421948} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.860957] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 888.861971] env[65522]: WARNING neutronclient.v2_0.client [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 888.861971] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.862148] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 888.862484] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 888.862853] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-218456f4-aa36-416a-8388-4e92addce32d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.871784] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 888.871784] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525eb310-0b4b-cdb9-1d3c-106b0c703950" [ 888.871784] env[65522]: _type = "Task" [ 888.871784] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.882505] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525eb310-0b4b-cdb9-1d3c-106b0c703950, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.907749] env[65522]: DEBUG oslo_concurrency.lockutils [None req-57c59541-e09c-47d9-b1d4-eccedafdc7f9 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "495fa98f-a9db-4214-87cc-d29209d3cb62" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 25.551s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 888.914850] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24934b3f-9079-4bbe-b277-0d7919f16108 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.928300] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16693912-6f98-4b67-8ce9-3a225bf4511f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.974751] env[65522]: DEBUG oslo_vmware.api [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Task: {'id': task-5114234, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194257} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.975553] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42022a88-48f2-484b-9a35-f27ae0e1af33 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.978395] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.978592] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 888.979275] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 888.979275] env[65522]: INFO nova.compute.manager [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Took 1.28 seconds to destroy the instance on the hypervisor. [ 888.979275] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 888.979492] env[65522]: DEBUG nova.compute.manager [-] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 888.979492] env[65522]: DEBUG nova.network.neutron [-] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 888.979735] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 888.980334] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.980571] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.996525] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a202213a-4531-44d6-9007-c6cd67088380 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.014213] env[65522]: DEBUG nova.compute.provider_tree [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.022200] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 889.091318] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ab0249-3be1-3834-58a2-115201f3f6ed, 'name': SearchDatastore_Task, 'duration_secs': 0.013429} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.092332] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ea508d6-7b66-42f5-94d3-2e71be50de6c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.101236] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 889.101236] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529836c6-8b75-554e-76c0-bfea9b5061ae" [ 889.101236] env[65522]: _type = "Task" [ 889.101236] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.114498] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529836c6-8b75-554e-76c0-bfea9b5061ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.158965] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114236, 'name': Rename_Task, 'duration_secs': 0.196231} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.159721] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 889.160052] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9ca3074-7d0b-4341-9ee8-d3f44e3b1339 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.167622] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 889.167622] env[65522]: value = "task-5114237" [ 889.167622] env[65522]: _type = "Task" [ 889.167622] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.170965] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114230, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.179942] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114237, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.253119] env[65522]: DEBUG oslo_vmware.api [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114231, 'name': PowerOnVM_Task, 'duration_secs': 0.561659} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.253474] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 889.253722] env[65522]: INFO nova.compute.manager [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Took 9.67 seconds to spawn the instance on the hypervisor. [ 889.253945] env[65522]: DEBUG nova.compute.manager [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 889.254858] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb7a5fa-d186-48c3-9a7c-465b8e09ba4d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.286857] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 889.287258] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72a1643d-52ad-4a9d-b131-0391529af24e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.295676] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 889.295676] env[65522]: value = "task-5114238" [ 889.295676] env[65522]: _type = "Task" [ 889.295676] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.308677] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114238, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.385298] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525eb310-0b4b-cdb9-1d3c-106b0c703950, 'name': SearchDatastore_Task, 'duration_secs': 0.031794} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.385673] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 889.386184] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 889.386282] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.517357] env[65522]: DEBUG nova.scheduler.client.report [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 889.618026] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529836c6-8b75-554e-76c0-bfea9b5061ae, 'name': SearchDatastore_Task, 'duration_secs': 0.031163} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.618026] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 889.618026] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 1eebb320-30e2-4d8b-b1fd-19e31fb35b77/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk. {{(pid=65522) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 889.618026] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 889.618026] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 889.618026] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dd73ac2d-4290-4ee0-a1b3-5e978f7a4232 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.618733] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-80fa1495-cfbe-49a0-a043-367c7a53e3e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.626716] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 889.626716] env[65522]: value = "task-5114239" [ 889.626716] env[65522]: _type = "Task" [ 889.626716] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.631496] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 889.631707] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 889.632915] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e315366-46c2-4030-9576-ab7076d468a0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.639209] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114239, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.643301] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 889.643301] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522caa3d-1e5a-884b-d0f4-1f5d98465d21" [ 889.643301] env[65522]: _type = "Task" [ 889.643301] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.653768] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522caa3d-1e5a-884b-d0f4-1f5d98465d21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.674065] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114230, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.014341} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.678192] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 9da312a0-c6eb-4903-a7d8-e178aa54d2ea/9da312a0-c6eb-4903-a7d8-e178aa54d2ea.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 889.678411] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.682534] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d8960e3-e72f-4351-9c59-11906ad0a871 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.683449] env[65522]: DEBUG nova.compute.manager [req-12fe576b-d7e7-4c16-b3e0-eeabd111386d req-033fbb90-dd44-4d40-9e7c-9b355756c07c service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Received event network-vif-deleted-990aae71-9e8e-4da6-9e57-5c9257d127a4 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 889.683584] env[65522]: INFO nova.compute.manager [req-12fe576b-d7e7-4c16-b3e0-eeabd111386d req-033fbb90-dd44-4d40-9e7c-9b355756c07c service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Neutron deleted interface 990aae71-9e8e-4da6-9e57-5c9257d127a4; detaching it from the instance and deleting it from the info cache [ 889.683770] env[65522]: DEBUG nova.network.neutron [req-12fe576b-d7e7-4c16-b3e0-eeabd111386d req-033fbb90-dd44-4d40-9e7c-9b355756c07c service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 889.692542] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114237, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.694378] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Waiting for the task: (returnval){ [ 889.694378] env[65522]: value = "task-5114240" [ 889.694378] env[65522]: _type = "Task" [ 889.694378] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.774525] env[65522]: INFO nova.compute.manager [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Took 31.25 seconds to build instance. [ 889.809378] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114238, 'name': PowerOffVM_Task, 'duration_secs': 0.196551} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.809689] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 889.809923] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance '5e227f6c-9f13-416f-8e6d-2f7d931619fd' progress to 17 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 889.955523] env[65522]: DEBUG nova.network.neutron [-] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 890.025221] env[65522]: DEBUG oslo_concurrency.lockutils [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.198s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 890.026407] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.521s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.029516] env[65522]: INFO nova.compute.claims [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 890.062795] env[65522]: INFO nova.scheduler.client.report [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Deleted allocations for instance 361e54c7-4309-48b3-ad3e-b2ebc783741c [ 890.144873] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114239, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.159543] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522caa3d-1e5a-884b-d0f4-1f5d98465d21, 'name': SearchDatastore_Task, 'duration_secs': 0.013008} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.160956] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d0d55d5-4cdc-4309-911f-e08d6a4a5e43 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.170667] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 890.170667] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ec9d3d-573d-68f9-09c4-36a75357131f" [ 890.170667] env[65522]: _type = "Task" [ 890.170667] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.189325] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ec9d3d-573d-68f9-09c4-36a75357131f, 'name': SearchDatastore_Task, 'duration_secs': 0.012537} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.192704] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-393e1412-9d36-479b-9804-3726b602d2e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.195028] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 890.195322] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] bb20d99f-fc49-47bf-a6b7-74ea966b71eb/bb20d99f-fc49-47bf-a6b7-74ea966b71eb.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 890.195614] env[65522]: DEBUG oslo_vmware.api [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114237, 'name': PowerOnVM_Task, 'duration_secs': 0.564266} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.196091] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21c7f7c1-8237-44ec-875a-9b6cf2fb370f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.198540] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 890.198744] env[65522]: INFO nova.compute.manager [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Took 7.90 seconds to spawn the instance on the hypervisor. [ 890.198921] env[65522]: DEBUG nova.compute.manager [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 890.202811] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-648444ba-5e43-464e-abfc-f77a0a151022 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.211454] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09113cf6-7393-4c69-8281-acda81594b43 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.232290] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114240, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100495} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.233478] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 890.233478] env[65522]: value = "task-5114241" [ 890.233478] env[65522]: _type = "Task" [ 890.233478] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.234351] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.235299] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-469b4623-8e41-4b1e-8418-7d9cfb1bdad5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.256374] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114241, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.265476] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 9da312a0-c6eb-4903-a7d8-e178aa54d2ea/9da312a0-c6eb-4903-a7d8-e178aa54d2ea.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.282279] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9e84d0d-46dd-403a-bdd8-80da0cfeb1c0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.297689] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7573a057-165a-4b51-8caa-7980294945bd tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "308330f8-1219-4c02-a129-22b29f0e33b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.785s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 890.299091] env[65522]: DEBUG nova.compute.manager [req-12fe576b-d7e7-4c16-b3e0-eeabd111386d req-033fbb90-dd44-4d40-9e7c-9b355756c07c service nova] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Detach interface failed, port_id=990aae71-9e8e-4da6-9e57-5c9257d127a4, reason: Instance 18568baa-83de-4b54-addf-dee66211c8d7 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 890.309022] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Waiting for the task: (returnval){ [ 890.309022] env[65522]: value = "task-5114242" [ 890.309022] env[65522]: _type = "Task" [ 890.309022] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.326675] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 890.327389] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 890.327389] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 890.327571] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 890.327854] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 890.328116] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 890.328367] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 890.328832] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 890.328832] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 890.328832] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 890.329269] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 890.335349] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114242, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.336028] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-409192e4-e245-48c0-99f8-469532198b2a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.358024] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 890.358024] env[65522]: value = "task-5114243" [ 890.358024] env[65522]: _type = "Task" [ 890.358024] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.368334] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114243, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.459031] env[65522]: INFO nova.compute.manager [-] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Took 1.48 seconds to deallocate network for instance. [ 890.573337] env[65522]: DEBUG oslo_concurrency.lockutils [None req-51d731f9-0abf-434f-9b77-bab0e3a2b4ba tempest-ListImageFiltersTestJSON-972265694 tempest-ListImageFiltersTestJSON-972265694-project-member] Lock "361e54c7-4309-48b3-ad3e-b2ebc783741c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.198s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 890.637669] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114239, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55964} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.637983] env[65522]: INFO nova.virt.vmwareapi.ds_util [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 1eebb320-30e2-4d8b-b1fd-19e31fb35b77/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk. [ 890.638785] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8a8a29-dae0-4042-8a3f-4089baf1e3f5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.667811] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] 1eebb320-30e2-4d8b-b1fd-19e31fb35b77/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.668112] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-767364cf-1936-4e3c-a794-b18dbf427bf1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.687139] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 890.687139] env[65522]: value = "task-5114244" [ 890.687139] env[65522]: _type = "Task" [ 890.687139] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.697186] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114244, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.745760] env[65522]: INFO nova.compute.manager [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Took 27.37 seconds to build instance. [ 890.751043] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114241, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.805837] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "308330f8-1219-4c02-a129-22b29f0e33b2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.806214] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "308330f8-1219-4c02-a129-22b29f0e33b2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.806371] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "308330f8-1219-4c02-a129-22b29f0e33b2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.806553] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "308330f8-1219-4c02-a129-22b29f0e33b2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.806732] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "308330f8-1219-4c02-a129-22b29f0e33b2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 890.810629] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "495fa98f-a9db-4214-87cc-d29209d3cb62" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.810927] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "495fa98f-a9db-4214-87cc-d29209d3cb62" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.811395] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "495fa98f-a9db-4214-87cc-d29209d3cb62-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.811546] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "495fa98f-a9db-4214-87cc-d29209d3cb62-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.811779] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "495fa98f-a9db-4214-87cc-d29209d3cb62-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 890.815633] env[65522]: INFO nova.compute.manager [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Terminating instance [ 890.827972] env[65522]: INFO nova.compute.manager [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Terminating instance [ 890.842027] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114242, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.871738] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114243, 'name': ReconfigVM_Task, 'duration_secs': 0.464966} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.872167] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance '5e227f6c-9f13-416f-8e6d-2f7d931619fd' progress to 33 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 890.965213] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 891.202600] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114244, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.249267] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58dabb81-fbf7-4ae3-ab28-c5781d8d3ba0 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "4264998b-949d-4b01-8832-fb1df6214403" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.890s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 891.258013] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114241, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.328070] env[65522]: DEBUG nova.compute.manager [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 891.328464] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 891.338266] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6541a38c-36fd-437f-9901-a77ded1c443f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.338670] env[65522]: DEBUG nova.compute.manager [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 891.338839] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 891.339161] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114242, 'name': ReconfigVM_Task, 'duration_secs': 0.733873} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.340289] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a058d85e-a130-4d8e-93c5-fb3f2d29ae46 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.343364] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 9da312a0-c6eb-4903-a7d8-e178aa54d2ea/9da312a0-c6eb-4903-a7d8-e178aa54d2ea.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.344793] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c316e224-9d0e-493d-b44c-9b956eea5377 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.359811] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 891.360235] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 891.360675] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Waiting for the task: (returnval){ [ 891.360675] env[65522]: value = "task-5114245" [ 891.360675] env[65522]: _type = "Task" [ 891.360675] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.360960] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8cf2ca28-ceef-4a14-9ea5-db4fbcf048fb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.364035] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80417935-70c6-4a89-b3c6-f5c7f2721b68 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.383509] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:01:50Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1d8f2c10-9623-4a70-b964-018e648fd761',id=39,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-811894183',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 891.383782] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 891.383946] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 891.388157] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 891.388432] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 891.388645] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 891.388897] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 891.389076] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 891.389276] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 891.389552] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 891.389745] env[65522]: DEBUG nova.virt.hardware [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 891.399631] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Reconfiguring VM instance instance-00000039 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 891.400759] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114245, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.413290] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c80df17-d376-4d5f-a246-d0b901f88104 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.432437] env[65522]: DEBUG oslo_vmware.api [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 891.432437] env[65522]: value = "task-5114246" [ 891.432437] env[65522]: _type = "Task" [ 891.432437] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.433019] env[65522]: DEBUG oslo_vmware.api [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 891.433019] env[65522]: value = "task-5114247" [ 891.433019] env[65522]: _type = "Task" [ 891.433019] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.433652] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 891.433998] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.447860] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 891.447860] env[65522]: value = "task-5114248" [ 891.447860] env[65522]: _type = "Task" [ 891.447860] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.465607] env[65522]: DEBUG oslo_vmware.api [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114246, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.466349] env[65522]: DEBUG oslo_concurrency.lockutils [None req-085b0a76-54d9-45f9-93c5-ec6babf4e8ea tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "4264998b-949d-4b01-8832-fb1df6214403" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 891.466349] env[65522]: DEBUG oslo_concurrency.lockutils [None req-085b0a76-54d9-45f9-93c5-ec6babf4e8ea tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "4264998b-949d-4b01-8832-fb1df6214403" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.466655] env[65522]: DEBUG nova.compute.manager [None req-085b0a76-54d9-45f9-93c5-ec6babf4e8ea tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 891.466907] env[65522]: DEBUG oslo_vmware.api [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114247, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.468943] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a8b30f-5da1-485b-8f93-0a6413a69563 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.482406] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114248, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.487788] env[65522]: DEBUG nova.compute.manager [None req-085b0a76-54d9-45f9-93c5-ec6babf4e8ea tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65522) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 891.489394] env[65522]: DEBUG nova.objects.instance [None req-085b0a76-54d9-45f9-93c5-ec6babf4e8ea tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lazy-loading 'flavor' on Instance uuid 4264998b-949d-4b01-8832-fb1df6214403 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.661191] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e50cfaed-44fd-43cd-97bf-47c798f837e0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.674065] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d462babc-ff4f-4de8-963a-e8be172504af {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.716351] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-685e46f5-0b54-444b-8016-796784189ed8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.727322] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49434c61-a3d6-4fa4-867d-183c40ed785a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.731658] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114244, 'name': ReconfigVM_Task, 'duration_secs': 0.744218} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.731955] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Reconfigured VM instance instance-0000002e to attach disk [datastore2] 1eebb320-30e2-4d8b-b1fd-19e31fb35b77/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.733266] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a064d199-1018-4c4a-b486-bca90d9e3e52 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.744541] env[65522]: DEBUG nova.compute.provider_tree [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.775242] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f9909c5-898c-475a-be53-68ee3ff995f1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.789378] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114241, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.496253} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.790035] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] bb20d99f-fc49-47bf-a6b7-74ea966b71eb/bb20d99f-fc49-47bf-a6b7-74ea966b71eb.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 891.790304] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 891.792687] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-61a73699-f4f5-4c84-a670-275cbb6d213a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.794859] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 891.794859] env[65522]: value = "task-5114249" [ 891.794859] env[65522]: _type = "Task" [ 891.794859] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.799295] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 891.799295] env[65522]: value = "task-5114250" [ 891.799295] env[65522]: _type = "Task" [ 891.799295] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.808503] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114249, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.817144] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114250, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.876091] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114245, 'name': Rename_Task, 'duration_secs': 0.314069} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.876380] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 891.876644] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa3274dd-083c-49ca-8f5b-d9006502c713 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.883335] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Waiting for the task: (returnval){ [ 891.883335] env[65522]: value = "task-5114251" [ 891.883335] env[65522]: _type = "Task" [ 891.883335] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.893135] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114251, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.949110] env[65522]: DEBUG nova.compute.manager [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 891.951867] env[65522]: DEBUG oslo_vmware.api [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114246, 'name': PowerOffVM_Task, 'duration_secs': 0.316703} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.955738] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.955848] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 891.956255] env[65522]: DEBUG oslo_vmware.api [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114247, 'name': PowerOffVM_Task, 'duration_secs': 0.316121} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.956611] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2c8b3819-a956-4de6-adfa-a1561a7aeea7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.961129] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.961326] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 891.961859] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-27b5d97f-69dc-4e39-943f-7a9f9f602ea3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.969587] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114248, 'name': ReconfigVM_Task, 'duration_secs': 0.262481} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.969867] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Reconfigured VM instance instance-00000039 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 891.970826] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06abf9d9-1c6d-4061-8027-bd485c7fa89c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.996526] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 5e227f6c-9f13-416f-8e6d-2f7d931619fd/5e227f6c-9f13-416f-8e6d-2f7d931619fd.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.000220] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ba830e4-8481-413b-b804-f643d1343ab0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.022223] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 892.022223] env[65522]: value = "task-5114254" [ 892.022223] env[65522]: _type = "Task" [ 892.022223] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.033536] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114254, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.063684] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 892.063684] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 892.063684] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Deleting the datastore file [datastore1] 308330f8-1219-4c02-a129-22b29f0e33b2 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 892.065413] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3ad830b-7369-4adb-8824-f0fd7a3e250f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.076310] env[65522]: DEBUG oslo_vmware.api [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for the task: (returnval){ [ 892.076310] env[65522]: value = "task-5114255" [ 892.076310] env[65522]: _type = "Task" [ 892.076310] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.081953] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 892.082196] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 892.082429] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleting the datastore file [datastore1] 495fa98f-a9db-4214-87cc-d29209d3cb62 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 892.083248] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d530e1f-4aa0-4dc4-962b-a58e8f136a16 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.089085] env[65522]: DEBUG oslo_vmware.api [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.094568] env[65522]: DEBUG oslo_vmware.api [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 892.094568] env[65522]: value = "task-5114256" [ 892.094568] env[65522]: _type = "Task" [ 892.094568] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.108294] env[65522]: DEBUG oslo_vmware.api [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114256, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.271993] env[65522]: DEBUG nova.scheduler.client.report [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 892.309119] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114249, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.314035] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114250, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070678} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.314376] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 892.315191] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25524ab2-ca5c-4102-97f6-02734b565426 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.339225] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] bb20d99f-fc49-47bf-a6b7-74ea966b71eb/bb20d99f-fc49-47bf-a6b7-74ea966b71eb.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.339597] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-637bbded-c9fb-466b-8be0-6d541749ca2c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.360389] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 892.360389] env[65522]: value = "task-5114257" [ 892.360389] env[65522]: _type = "Task" [ 892.360389] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.374070] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114257, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.394430] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114251, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.478673] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 892.516538] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-085b0a76-54d9-45f9-93c5-ec6babf4e8ea tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 892.516903] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5f6052f-fc0d-4fb1-ba6e-f35e9bcf52ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.529233] env[65522]: DEBUG oslo_vmware.api [None req-085b0a76-54d9-45f9-93c5-ec6babf4e8ea tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 892.529233] env[65522]: value = "task-5114258" [ 892.529233] env[65522]: _type = "Task" [ 892.529233] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.536912] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114254, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.542274] env[65522]: DEBUG oslo_vmware.api [None req-085b0a76-54d9-45f9-93c5-ec6babf4e8ea tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114258, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.586918] env[65522]: DEBUG oslo_vmware.api [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Task: {'id': task-5114255, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196589} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.587289] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.587524] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 892.587744] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 892.587974] env[65522]: INFO nova.compute.manager [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Took 1.26 seconds to destroy the instance on the hypervisor. [ 892.588259] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 892.588459] env[65522]: DEBUG nova.compute.manager [-] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 892.588549] env[65522]: DEBUG nova.network.neutron [-] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 892.588799] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.589390] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 892.589727] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 892.606514] env[65522]: DEBUG oslo_vmware.api [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114256, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.206516} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.606775] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.606987] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 892.607152] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 892.607335] env[65522]: INFO nova.compute.manager [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Took 1.27 seconds to destroy the instance on the hypervisor. [ 892.607736] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 892.608789] env[65522]: DEBUG nova.compute.manager [-] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 892.608789] env[65522]: DEBUG nova.network.neutron [-] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 892.608789] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.608789] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 892.609226] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 892.679699] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.683060] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.780261] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.754s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 892.781601] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 20.955s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 892.809056] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114249, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.874741] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114257, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.897684] env[65522]: DEBUG oslo_vmware.api [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114251, 'name': PowerOnVM_Task, 'duration_secs': 0.632667} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.898098] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 892.898387] env[65522]: INFO nova.compute.manager [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Took 8.42 seconds to spawn the instance on the hypervisor. [ 892.898647] env[65522]: DEBUG nova.compute.manager [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 892.899744] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b04bf7-af8a-423a-8e7d-f566c4e46e35 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.036314] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114254, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.045776] env[65522]: DEBUG oslo_vmware.api [None req-085b0a76-54d9-45f9-93c5-ec6babf4e8ea tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114258, 'name': PowerOffVM_Task, 'duration_secs': 0.332474} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.046072] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-085b0a76-54d9-45f9-93c5-ec6babf4e8ea tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 893.046284] env[65522]: DEBUG nova.compute.manager [None req-085b0a76-54d9-45f9-93c5-ec6babf4e8ea tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 893.047172] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c7d26ae-9787-40e8-9575-0327e5326fa1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.076100] env[65522]: DEBUG nova.compute.manager [req-0a24219f-938a-44f0-863b-00b0a808158a req-a00f4b0e-ef52-456e-b950-c78bd79139a1 service nova] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Received event network-vif-deleted-eba45f53-1f29-46b1-8285-33b5aacbcc43 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 893.076183] env[65522]: INFO nova.compute.manager [req-0a24219f-938a-44f0-863b-00b0a808158a req-a00f4b0e-ef52-456e-b950-c78bd79139a1 service nova] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Neutron deleted interface eba45f53-1f29-46b1-8285-33b5aacbcc43; detaching it from the instance and deleting it from the info cache [ 893.076357] env[65522]: DEBUG nova.network.neutron [req-0a24219f-938a-44f0-863b-00b0a808158a req-a00f4b0e-ef52-456e-b950-c78bd79139a1 service nova] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 893.097980] env[65522]: DEBUG nova.compute.manager [req-4b8756ac-2551-4e8e-a177-2f0ca86a7bd4 req-132b0e08-25f8-4f4d-acf7-80819f2264e9 service nova] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Received event network-vif-deleted-9446c1bb-b443-4a52-879d-5fa6cb83f0dc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 893.098216] env[65522]: INFO nova.compute.manager [req-4b8756ac-2551-4e8e-a177-2f0ca86a7bd4 req-132b0e08-25f8-4f4d-acf7-80819f2264e9 service nova] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Neutron deleted interface 9446c1bb-b443-4a52-879d-5fa6cb83f0dc; detaching it from the instance and deleting it from the info cache [ 893.098387] env[65522]: DEBUG nova.network.neutron [req-4b8756ac-2551-4e8e-a177-2f0ca86a7bd4 req-132b0e08-25f8-4f4d-acf7-80819f2264e9 service nova] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 893.284894] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Acquiring lock "14a6ec0f-63a9-46d9-9d21-3c4aaf2645dc" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 893.285194] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Lock "14a6ec0f-63a9-46d9-9d21-3c4aaf2645dc" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 893.313355] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114249, 'name': ReconfigVM_Task, 'duration_secs': 1.10912} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.314354] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.314657] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-643acddc-bf8f-45db-b306-c95c4d3ac5a1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.323665] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 893.323665] env[65522]: value = "task-5114259" [ 893.323665] env[65522]: _type = "Task" [ 893.323665] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.332544] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114259, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.371266] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114257, 'name': ReconfigVM_Task, 'duration_secs': 0.708411} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.371564] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Reconfigured VM instance instance-00000029 to attach disk [datastore2] bb20d99f-fc49-47bf-a6b7-74ea966b71eb/bb20d99f-fc49-47bf-a6b7-74ea966b71eb.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.372243] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eb9b3767-f1f5-40be-b42d-ce178e74904b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.379264] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 893.379264] env[65522]: value = "task-5114260" [ 893.379264] env[65522]: _type = "Task" [ 893.379264] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.389881] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114260, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.418432] env[65522]: INFO nova.compute.manager [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Took 28.52 seconds to build instance. [ 893.465074] env[65522]: DEBUG nova.network.neutron [-] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 893.469444] env[65522]: DEBUG nova.network.neutron [-] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 893.534665] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114254, 'name': ReconfigVM_Task, 'duration_secs': 1.021185} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.534985] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 5e227f6c-9f13-416f-8e6d-2f7d931619fd/5e227f6c-9f13-416f-8e6d-2f7d931619fd.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.535367] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance '5e227f6c-9f13-416f-8e6d-2f7d931619fd' progress to 50 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 893.560398] env[65522]: DEBUG oslo_concurrency.lockutils [None req-085b0a76-54d9-45f9-93c5-ec6babf4e8ea tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "4264998b-949d-4b01-8832-fb1df6214403" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.094s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 893.579862] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f243593e-ea2c-425e-bfce-f7d742429698 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.590760] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602ced12-fc2f-4722-ae28-2f5b569e602d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.608395] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-79776449-479d-41b3-9151-4cc98bdc1fc5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.622263] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2bbd8ca-8e14-4533-a8de-4dc3276d9960 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.653213] env[65522]: DEBUG nova.compute.manager [req-0a24219f-938a-44f0-863b-00b0a808158a req-a00f4b0e-ef52-456e-b950-c78bd79139a1 service nova] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Detach interface failed, port_id=eba45f53-1f29-46b1-8285-33b5aacbcc43, reason: Instance 308330f8-1219-4c02-a129-22b29f0e33b2 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 893.686334] env[65522]: DEBUG nova.compute.manager [req-4b8756ac-2551-4e8e-a177-2f0ca86a7bd4 req-132b0e08-25f8-4f4d-acf7-80819f2264e9 service nova] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Detach interface failed, port_id=9446c1bb-b443-4a52-879d-5fa6cb83f0dc, reason: Instance 495fa98f-a9db-4214-87cc-d29209d3cb62 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 893.789552] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Lock "14a6ec0f-63a9-46d9-9d21-3c4aaf2645dc" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.504s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 893.790275] env[65522]: DEBUG nova.compute.manager [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 893.802770] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Applying migration context for instance 5e227f6c-9f13-416f-8e6d-2f7d931619fd as it has an incoming, in-progress migration ed83be76-1d3a-4a25-85ce-9cf749f05abc. Migration status is migrating {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 893.805441] env[65522]: INFO nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating resource usage from migration ed83be76-1d3a-4a25-85ce-9cf749f05abc [ 893.832820] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7c6f8218-602d-44f3-8012-de5a96972785 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.833018] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance e5d2cc50-4ee6-4e64-9b52-888968a717ca is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 893.833165] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance ca964440-5375-4aff-8b45-96fbe829dd16 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.833281] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 87a3c63d-794d-44ab-bad6-65c323d72ae7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.833394] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 68d8d59b-4fc8-4993-8c9e-31b2329c88dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.833518] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 75c8848b-aea9-43f1-8697-9224050d1fef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.833632] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 42a08cfb-d865-4967-a086-370a2ca98b7d actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.834403] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance d5e8dd05-dc3c-4831-b4b0-ac100360f3e7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.835690] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance bb20d99f-fc49-47bf-a6b7-74ea966b71eb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.835690] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 24f520ec-6a56-4f17-9ae4-2c856bf99582 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.835690] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 1eebb320-30e2-4d8b-b1fd-19e31fb35b77 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.835690] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 05319f69-2ac7-4649-a90a-3542d87a3e82 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.836049] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 18568baa-83de-4b54-addf-dee66211c8d7 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 893.836108] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 0b48b747-149c-4163-bdd0-7d0b07ddacfd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.836248] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 77b3ca48-d2c6-4743-a44a-53b9aa84662a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 893.836371] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7617189c-a902-42e7-8165-0e7c4a1de06d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.836482] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 495fa98f-a9db-4214-87cc-d29209d3cb62 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.836604] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 26432ad6-610e-49c2-8699-f1e7e128e5b8 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 893.836720] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 8eda6816-6781-4ad9-9210-11fa623f24b0 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 893.836848] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 179a37ed-1176-4082-8c68-1be6abd3fd9d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 893.836944] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 893.837068] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 9df1a51a-2811-4486-a4c6-58d618f2ae7d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.837181] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 1e7b8237-34ea-479e-b5a7-b1846661d61d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.837292] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 308330f8-1219-4c02-a129-22b29f0e33b2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.837396] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 4264998b-949d-4b01-8832-fb1df6214403 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.837502] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 9da312a0-c6eb-4903-a7d8-e178aa54d2ea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.837608] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Migration ed83be76-1d3a-4a25-85ce-9cf749f05abc is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 893.837713] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 5e227f6c-9f13-416f-8e6d-2f7d931619fd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.837833] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 55ab4771-9908-4640-a142-3cb40c0c6ee5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 893.847151] env[65522]: DEBUG oslo_vmware.api [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114259, 'name': PowerOnVM_Task, 'duration_secs': 0.466328} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.848112] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 893.851095] env[65522]: DEBUG nova.compute.manager [None req-5ed65a70-5c34-4bbb-a7fc-6b646602e207 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 893.852305] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9653fba8-fbd1-4492-860d-25def2c9eae6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.889726] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114260, 'name': Rename_Task, 'duration_secs': 0.19382} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.890445] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.890722] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5656bab6-2093-469b-9f16-882759fbefa9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.899547] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 893.899547] env[65522]: value = "task-5114261" [ 893.899547] env[65522]: _type = "Task" [ 893.899547] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.910239] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114261, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.920438] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c53fbc03-39be-4959-8df4-5a2f5d1dafea tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Lock "9da312a0-c6eb-4903-a7d8-e178aa54d2ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.046s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 893.969202] env[65522]: INFO nova.compute.manager [-] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Took 1.36 seconds to deallocate network for instance. [ 893.971278] env[65522]: INFO nova.compute.manager [-] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Took 1.38 seconds to deallocate network for instance. [ 894.043179] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a55e14c-cf50-4226-8766-cb69aab17c85 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.066568] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6d9853-e53a-4101-9ded-e3df26842b74 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.087165] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance '5e227f6c-9f13-416f-8e6d-2f7d931619fd' progress to 67 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 894.194625] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Acquiring lock "9da312a0-c6eb-4903-a7d8-e178aa54d2ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.194921] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Lock "9da312a0-c6eb-4903-a7d8-e178aa54d2ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.195164] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Acquiring lock "9da312a0-c6eb-4903-a7d8-e178aa54d2ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.195353] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Lock "9da312a0-c6eb-4903-a7d8-e178aa54d2ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.195523] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Lock "9da312a0-c6eb-4903-a7d8-e178aa54d2ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 894.198348] env[65522]: INFO nova.compute.manager [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Terminating instance [ 894.296731] env[65522]: DEBUG nova.compute.utils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 894.300087] env[65522]: DEBUG nova.compute.manager [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 894.300327] env[65522]: DEBUG nova.network.neutron [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 894.301859] env[65522]: WARNING neutronclient.v2_0.client [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 894.301859] env[65522]: WARNING neutronclient.v2_0.client [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 894.301859] env[65522]: WARNING openstack [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 894.302362] env[65522]: WARNING openstack [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 894.342262] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 9c39b722-56ff-44fa-8f66-3e3432645a68 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 894.598038] env[65522]: DEBUG nova.policy [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a5907199463d4f02bceacd062d1ca6f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e096164b91f8483994f1dfefbca5fdff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 894.605654] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.605654] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.607232] env[65522]: WARNING neutronclient.v2_0.client [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 894.623219] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114261, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.703889] env[65522]: DEBUG nova.compute.manager [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 894.704325] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 894.705220] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e316380-b2d4-4db9-87c3-c1420140e371 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.714214] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 894.714599] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c7e91c44-d4f3-46bc-92d5-d5a37ca132d5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.722370] env[65522]: DEBUG oslo_vmware.api [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Waiting for the task: (returnval){ [ 894.722370] env[65522]: value = "task-5114262" [ 894.722370] env[65522]: _type = "Task" [ 894.722370] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.731673] env[65522]: DEBUG oslo_vmware.api [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114262, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.788616] env[65522]: DEBUG nova.network.neutron [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Port 3f79042c-23dd-4fc0-b3cf-1e2884edabc1 binding to destination host cpu-1 is already ACTIVE {{(pid=65522) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 894.811414] env[65522]: DEBUG nova.compute.manager [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 894.846642] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance a2a5fb2d-62e0-4809-a01f-f1df66dca58f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 894.846983] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 22 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 894.847181] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4928MB phys_disk=100GB used_disk=22GB total_vcpus=48 used_vcpus=22 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '21', 'num_vm_active': '17', 'num_task_None': '14', 'num_os_type_None': '21', 'num_proj_fa11b46d9fe144f391233e6eb9c819d7': '1', 'io_workload': '4', 'num_proj_9b8112ba49034e6fabcb3fbbd46edf41': '3', 'num_vm_rescued': '1', 'num_proj_eb9a9e5da1194ab7927f72d21d1484b0': '1', 'num_proj_503adb79a1f74f1b98e1f06ec54f1584': '2', 'num_proj_37fc01876567476f9b93d765b2cfddc8': '1', 'num_proj_6d64b94285334dd6b26928cabedbd67c': '1', 'num_vm_error': '1', 'num_task_rebuild_spawning': '1', 'num_proj_b5c14a856d454048856539895f220c2c': '4', 'num_task_deleting': '2', 'num_proj_6ecccb656b0d4c96b40b200cdcddbad5': '1', 'num_task_rescuing': '1', 'num_proj_234a226c969b4965bfa2c411bb4c8d90': '1', 'num_proj_91597233ae9c44c094f4c32d90332fa6': '1', 'num_task_resize_migrating': '1', 'num_proj_068c2387de8c406194d9b1762c7292a7': '1', 'num_proj_e2bd38f300424db78568e0b8d9982f27': '1', 'num_task_powering-off': '1', 'num_proj_3ea0fab7de6c47eeaf3bdde413ae3901': '1', 'num_vm_building': '2', 'num_task_spawning': '1', 'num_proj_cb18797424f3489ab6747cb859d7ff23': '1', 'num_proj_e096164b91f8483994f1dfefbca5fdff': '1'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 894.999351] env[65522]: DEBUG nova.network.neutron [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Successfully created port: 4463f170-2656-4332-b1a3-0b0a491984da {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 895.118219] env[65522]: DEBUG oslo_vmware.api [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114261, 'name': PowerOnVM_Task, 'duration_secs': 1.17354} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.118511] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.118695] env[65522]: DEBUG nova.compute.manager [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 895.119509] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5df156b-4e23-4e90-bb3d-1f506b3a165c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.236352] env[65522]: DEBUG oslo_vmware.api [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114262, 'name': PowerOffVM_Task, 'duration_secs': 0.194207} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.236736] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.236967] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 895.237210] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6396449f-0dcb-4fa8-bf64-c0a4689ce9f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.311896] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.312132] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.312344] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Deleting the datastore file [datastore1] 9da312a0-c6eb-4903-a7d8-e178aa54d2ea {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.312616] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5c3b6fd-ee12-41b8-9588-1ec9bf5e5ead {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.327355] env[65522]: DEBUG oslo_vmware.api [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Waiting for the task: (returnval){ [ 895.327355] env[65522]: value = "task-5114264" [ 895.327355] env[65522]: _type = "Task" [ 895.327355] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.339570] env[65522]: DEBUG oslo_vmware.api [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114264, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.344319] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80073c11-0525-4a0e-9615-e3ed8b14b1d2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.354418] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a44323-9a47-40ad-8dea-40d822422a71 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.390156] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65c4f51d-2d02-4338-a697-50dfc2ed6ce4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.399209] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a8654c7-c429-4b96-a299-f98f6c0456e2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.417176] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.636779] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 895.672126] env[65522]: DEBUG nova.compute.manager [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 895.673096] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fd5910-cad2-47ea-bb31-a78a24e3d54f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.820024] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 895.820024] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 895.820024] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 895.823420] env[65522]: DEBUG nova.compute.manager [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 895.838127] env[65522]: DEBUG oslo_vmware.api [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Task: {'id': task-5114264, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199135} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.839036] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.839036] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 895.839036] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 895.839291] env[65522]: INFO nova.compute.manager [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Took 1.14 seconds to destroy the instance on the hypervisor. [ 895.839468] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 895.839740] env[65522]: DEBUG nova.compute.manager [-] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 895.839848] env[65522]: DEBUG nova.network.neutron [-] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 895.840168] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.840783] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.841149] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.857662] env[65522]: DEBUG nova.virt.hardware [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 895.857895] env[65522]: DEBUG nova.virt.hardware [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 895.858117] env[65522]: DEBUG nova.virt.hardware [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 895.858329] env[65522]: DEBUG nova.virt.hardware [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 895.858493] env[65522]: DEBUG nova.virt.hardware [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 895.858650] env[65522]: DEBUG nova.virt.hardware [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 895.858865] env[65522]: DEBUG nova.virt.hardware [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.859068] env[65522]: DEBUG nova.virt.hardware [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 895.859264] env[65522]: DEBUG nova.virt.hardware [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 895.859441] env[65522]: DEBUG nova.virt.hardware [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 895.859617] env[65522]: DEBUG nova.virt.hardware [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 895.861333] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231e957a-fcb4-4767-9e34-fba20c6192ac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.871494] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0515b7f6-77f1-46eb-ab79-a84129a57c2f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.921297] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 896.185694] env[65522]: INFO nova.compute.manager [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] instance snapshotting [ 896.185869] env[65522]: WARNING nova.compute.manager [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 896.188902] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e46bcdb-248e-46cb-89ef-96d7ee5a8733 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.211616] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2465bed-3278-43e0-8ffe-c776a700b33f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.426306] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 896.426644] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.645s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 896.426938] env[65522]: DEBUG oslo_concurrency.lockutils [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.851s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 896.427132] env[65522]: DEBUG nova.objects.instance [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 896.429949] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 896.430138] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Cleaning up deleted instances {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11933}} [ 896.546562] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 896.609133] env[65522]: DEBUG nova.network.neutron [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Successfully updated port: 4463f170-2656-4332-b1a3-0b0a491984da {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 896.730279] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 896.730279] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8525e83c-66d0-48b5-8557-0fa3ed960928 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.741794] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 896.741794] env[65522]: value = "task-5114265" [ 896.741794] env[65522]: _type = "Task" [ 896.741794] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.752244] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114265, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.826263] env[65522]: WARNING neutronclient.v2_0.client [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 896.909067] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.909358] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 896.909484] env[65522]: DEBUG nova.network.neutron [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 896.958403] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] There are 38 instances to clean {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11942}} [ 896.958540] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 880390f5-00f1-4dea-9345-29c06289b340] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 896.963238] env[65522]: DEBUG oslo_concurrency.lockutils [None req-025703de-895d-44df-8710-7ac0d486006a tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.536s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 896.964637] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.120s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 896.966020] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 896.967435] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.676s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 896.967746] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 896.970133] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.847s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 896.971908] env[65522]: INFO nova.compute.claims [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 897.037017] env[65522]: INFO nova.scheduler.client.report [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Deleted allocations for instance 8eda6816-6781-4ad9-9210-11fa623f24b0 [ 897.044589] env[65522]: INFO nova.scheduler.client.report [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Deleted allocations for instance 77b3ca48-d2c6-4743-a44a-53b9aa84662a [ 897.085070] env[65522]: INFO nova.compute.manager [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Unrescuing [ 897.085132] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.085248] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquired lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 897.085414] env[65522]: DEBUG nova.network.neutron [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 897.111792] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Acquiring lock "refresh_cache-55ab4771-9908-4640-a142-3cb40c0c6ee5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.112825] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Acquired lock "refresh_cache-55ab4771-9908-4640-a142-3cb40c0c6ee5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 897.112970] env[65522]: DEBUG nova.network.neutron [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 897.181970] env[65522]: DEBUG nova.compute.manager [req-3e1266ae-7813-433f-bfc1-261f97c29ec3 req-f2a22578-645d-4e1d-9679-e18799962c6e service nova] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Received event network-vif-plugged-4463f170-2656-4332-b1a3-0b0a491984da {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 897.181970] env[65522]: DEBUG oslo_concurrency.lockutils [req-3e1266ae-7813-433f-bfc1-261f97c29ec3 req-f2a22578-645d-4e1d-9679-e18799962c6e service nova] Acquiring lock "55ab4771-9908-4640-a142-3cb40c0c6ee5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 897.182839] env[65522]: DEBUG oslo_concurrency.lockutils [req-3e1266ae-7813-433f-bfc1-261f97c29ec3 req-f2a22578-645d-4e1d-9679-e18799962c6e service nova] Lock "55ab4771-9908-4640-a142-3cb40c0c6ee5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 897.182839] env[65522]: DEBUG oslo_concurrency.lockutils [req-3e1266ae-7813-433f-bfc1-261f97c29ec3 req-f2a22578-645d-4e1d-9679-e18799962c6e service nova] Lock "55ab4771-9908-4640-a142-3cb40c0c6ee5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 897.182839] env[65522]: DEBUG nova.compute.manager [req-3e1266ae-7813-433f-bfc1-261f97c29ec3 req-f2a22578-645d-4e1d-9679-e18799962c6e service nova] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] No waiting events found dispatching network-vif-plugged-4463f170-2656-4332-b1a3-0b0a491984da {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 897.182839] env[65522]: WARNING nova.compute.manager [req-3e1266ae-7813-433f-bfc1-261f97c29ec3 req-f2a22578-645d-4e1d-9679-e18799962c6e service nova] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Received unexpected event network-vif-plugged-4463f170-2656-4332-b1a3-0b0a491984da for instance with vm_state building and task_state spawning. [ 897.256091] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114265, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.282225] env[65522]: DEBUG nova.network.neutron [-] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 897.404864] env[65522]: DEBUG nova.compute.manager [req-3c36cc37-fdf6-4651-accc-dfb0314bd3e2 req-cde53a6d-bede-459f-b88c-7a77fedb116b service nova] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Received event network-vif-deleted-237a8db4-69b6-49b1-8828-e7aa07977495 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 897.416944] env[65522]: WARNING neutronclient.v2_0.client [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 897.418412] env[65522]: WARNING openstack [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.418556] env[65522]: WARNING openstack [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.477364] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 1be995d0-3141-4dcd-ad9a-278d9874e457] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 897.560436] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0857f724-755e-495b-bce0-6f7214e18aaf tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "77b3ca48-d2c6-4743-a44a-53b9aa84662a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.160s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 897.561686] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc6c2631-fc09-44f8-ae36-09a4009f12a1 tempest-ServerDiagnosticsNegativeTest-80843020 tempest-ServerDiagnosticsNegativeTest-80843020-project-member] Lock "8eda6816-6781-4ad9-9210-11fa623f24b0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.898s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 897.571830] env[65522]: WARNING openstack [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.572242] env[65522]: WARNING openstack [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.590156] env[65522]: WARNING neutronclient.v2_0.client [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 897.590156] env[65522]: WARNING openstack [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.590156] env[65522]: WARNING openstack [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.619334] env[65522]: WARNING openstack [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.619987] env[65522]: WARNING openstack [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.703609] env[65522]: DEBUG nova.network.neutron [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 897.736032] env[65522]: WARNING neutronclient.v2_0.client [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 897.736032] env[65522]: WARNING openstack [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.736032] env[65522]: WARNING openstack [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.753246] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114265, 'name': CreateSnapshot_Task, 'duration_secs': 0.794746} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.753733] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 897.754324] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbbc73b5-e571-459e-93e1-e47051d34601 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.787987] env[65522]: WARNING openstack [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.787987] env[65522]: WARNING openstack [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.793238] env[65522]: INFO nova.compute.manager [-] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Took 1.95 seconds to deallocate network for instance. [ 897.933213] env[65522]: WARNING openstack [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.933576] env[65522]: WARNING openstack [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.980893] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: cdee3a32-ad51-4364-8440-4744e8df322f] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 898.037576] env[65522]: DEBUG nova.network.neutron [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance_info_cache with network_info: [{"id": "3f79042c-23dd-4fc0-b3cf-1e2884edabc1", "address": "fa:16:3e:f7:f5:9e", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.131", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f79042c-23", "ovs_interfaceid": "3f79042c-23dd-4fc0-b3cf-1e2884edabc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 898.118175] env[65522]: WARNING neutronclient.v2_0.client [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 898.118844] env[65522]: WARNING openstack [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 898.119449] env[65522]: WARNING openstack [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 898.205901] env[65522]: WARNING neutronclient.v2_0.client [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 898.206986] env[65522]: WARNING openstack [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 898.207094] env[65522]: WARNING openstack [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 898.278934] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 898.286405] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-39d941b1-297d-4824-aa24-d35f32f4386d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.299442] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 898.299442] env[65522]: value = "task-5114266" [ 898.299442] env[65522]: _type = "Task" [ 898.299442] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.306043] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 898.315714] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114266, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.333125] env[65522]: DEBUG nova.network.neutron [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Updating instance_info_cache with network_info: [{"id": "4463f170-2656-4332-b1a3-0b0a491984da", "address": "fa:16:3e:28:f6:07", "network": {"id": "d1672958-02a6-48dd-98df-2663748a2aa1", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-384169563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e096164b91f8483994f1dfefbca5fdff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d88b750a-0e7d-4f16-8bd5-8e6d5743b720", "external-id": "nsx-vlan-transportzone-715", "segmentation_id": 715, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4463f170-26", "ovs_interfaceid": "4463f170-2656-4332-b1a3-0b0a491984da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 898.372766] env[65522]: DEBUG nova.network.neutron [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Updating instance_info_cache with network_info: [{"id": "e12c2307-4963-4a5a-af0f-19369217f430", "address": "fa:16:3e:9f:b8:7d", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape12c2307-49", "ovs_interfaceid": "e12c2307-4963-4a5a-af0f-19369217f430", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 898.372766] env[65522]: INFO nova.compute.manager [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Rebuilding instance [ 898.436543] env[65522]: DEBUG nova.compute.manager [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 898.437601] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d082d9-6422-4da2-82fb-442d86040e07 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.485424] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 9fa6e208-f913-4ecc-8548-ed6503fe6273] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 898.544489] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Releasing lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 898.587741] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d8db0c8-7fb9-4488-8d78-da70d4ef30ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.599253] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2934c209-0a19-461a-9882-58f9d8e0b07e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.643983] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7835c70-9c23-49c7-a58f-1834588e8f81 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.652130] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1054703-0a30-4a2e-ba36-b76ab650ff61 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.672149] env[65522]: DEBUG nova.compute.provider_tree [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.815030] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114266, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.835840] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Releasing lock "refresh_cache-55ab4771-9908-4640-a142-3cb40c0c6ee5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 898.836278] env[65522]: DEBUG nova.compute.manager [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Instance network_info: |[{"id": "4463f170-2656-4332-b1a3-0b0a491984da", "address": "fa:16:3e:28:f6:07", "network": {"id": "d1672958-02a6-48dd-98df-2663748a2aa1", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-384169563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e096164b91f8483994f1dfefbca5fdff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d88b750a-0e7d-4f16-8bd5-8e6d5743b720", "external-id": "nsx-vlan-transportzone-715", "segmentation_id": 715, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4463f170-26", "ovs_interfaceid": "4463f170-2656-4332-b1a3-0b0a491984da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 898.836774] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:f6:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd88b750a-0e7d-4f16-8bd5-8e6d5743b720', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4463f170-2656-4332-b1a3-0b0a491984da', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 898.846062] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Creating folder: Project (e096164b91f8483994f1dfefbca5fdff). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 898.846428] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d9feb158-8470-49f4-bbe8-50eb842e24ed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.863099] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Created folder: Project (e096164b91f8483994f1dfefbca5fdff) in parent group-v994660. [ 898.863099] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Creating folder: Instances. Parent ref: group-v994838. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 898.863099] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-523b6eec-6c92-4cfd-84ed-0e9fd048c856 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.876856] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Releasing lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 898.877539] env[65522]: DEBUG nova.objects.instance [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lazy-loading 'flavor' on Instance uuid 1eebb320-30e2-4d8b-b1fd-19e31fb35b77 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.879000] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Created folder: Instances in parent group-v994838. [ 898.879274] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 898.879700] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 898.880627] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4f0b51c-9fad-4d33-9d38-408dc515544c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.903963] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 898.903963] env[65522]: value = "task-5114269" [ 898.903963] env[65522]: _type = "Task" [ 898.903963] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.913941] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114269, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.994305] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: d2fdffb8-76d1-4b74-819f-25eb29ca91ce] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 899.072871] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece29c2a-7a7f-4f70-af73-36233d039c8e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.103512] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99fb6d7-f803-4adc-9c77-688e67dc63e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.114841] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance '5e227f6c-9f13-416f-8e6d-2f7d931619fd' progress to 83 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 899.174232] env[65522]: DEBUG nova.scheduler.client.report [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 899.274946] env[65522]: DEBUG nova.compute.manager [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Received event network-changed-4463f170-2656-4332-b1a3-0b0a491984da {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 899.275099] env[65522]: DEBUG nova.compute.manager [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Refreshing instance network info cache due to event network-changed-4463f170-2656-4332-b1a3-0b0a491984da. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 899.275324] env[65522]: DEBUG oslo_concurrency.lockutils [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] Acquiring lock "refresh_cache-55ab4771-9908-4640-a142-3cb40c0c6ee5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.275467] env[65522]: DEBUG oslo_concurrency.lockutils [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] Acquired lock "refresh_cache-55ab4771-9908-4640-a142-3cb40c0c6ee5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 899.275622] env[65522]: DEBUG nova.network.neutron [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Refreshing network info cache for port 4463f170-2656-4332-b1a3-0b0a491984da {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 899.311935] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114266, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.386272] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9314732e-4b08-48ff-a713-aa8619f641b4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.414726] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.418796] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6bb749f2-e3e1-40ca-bca0-9c019eea09ba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.427506] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114269, 'name': CreateVM_Task, 'duration_secs': 0.370571} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.428922] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 899.429468] env[65522]: DEBUG oslo_vmware.api [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 899.429468] env[65522]: value = "task-5114270" [ 899.429468] env[65522]: _type = "Task" [ 899.429468] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.429990] env[65522]: WARNING neutronclient.v2_0.client [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 899.430394] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.430553] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 899.430929] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 899.432189] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d9855a9-84f4-48c9-91e4-d23e8afae474 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.441686] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Waiting for the task: (returnval){ [ 899.441686] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52528852-996f-f680-161a-88532e72e2fd" [ 899.441686] env[65522]: _type = "Task" [ 899.441686] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.446847] env[65522]: DEBUG oslo_vmware.api [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114270, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.456843] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52528852-996f-f680-161a-88532e72e2fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.459696] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.460017] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5703214f-19f8-4b15-baee-9a9799212130 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.468817] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 899.468817] env[65522]: value = "task-5114271" [ 899.468817] env[65522]: _type = "Task" [ 899.468817] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.481335] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114271, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.498834] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 7693f017-69b3-4ec4-ae3f-7e65ed34a471] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 899.626840] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 899.627342] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3a90b3c4-ddff-45e8-989b-fc21fa5c1cad {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.638156] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 899.638156] env[65522]: value = "task-5114272" [ 899.638156] env[65522]: _type = "Task" [ 899.638156] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.657020] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114272, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.679825] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.710s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.680483] env[65522]: DEBUG nova.compute.manager [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 899.684243] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.098s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.684474] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.686813] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.658s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.687019] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.688918] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.936s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.689113] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.691425] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.579s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.691602] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.694851] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.728s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.694851] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.695409] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.217s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.697171] env[65522]: INFO nova.compute.claims [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 899.760516] env[65522]: INFO nova.scheduler.client.report [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Deleted allocations for instance 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e [ 899.765019] env[65522]: INFO nova.scheduler.client.report [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Deleted allocations for instance 26432ad6-610e-49c2-8699-f1e7e128e5b8 [ 899.778934] env[65522]: WARNING neutronclient.v2_0.client [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 899.780201] env[65522]: WARNING openstack [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 899.780640] env[65522]: WARNING openstack [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 899.788911] env[65522]: INFO nova.scheduler.client.report [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Deleted allocations for instance 179a37ed-1176-4082-8c68-1be6abd3fd9d [ 899.793025] env[65522]: INFO nova.scheduler.client.report [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Deleted allocations for instance 18568baa-83de-4b54-addf-dee66211c8d7 [ 899.806879] env[65522]: INFO nova.scheduler.client.report [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Deleted allocations for instance e5d2cc50-4ee6-4e64-9b52-888968a717ca [ 899.823905] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114266, 'name': CloneVM_Task} progress is 95%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.948636] env[65522]: DEBUG oslo_vmware.api [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114270, 'name': PowerOffVM_Task, 'duration_secs': 0.444735} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.952669] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 899.958299] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Reconfiguring VM instance instance-0000002e to detach disk 2002 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 899.958665] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77340f30-b61e-429c-ad82-d855b51a773f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.982047] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52528852-996f-f680-161a-88532e72e2fd, 'name': SearchDatastore_Task, 'duration_secs': 0.015438} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.984342] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 899.984745] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 899.985017] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.985173] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 899.985405] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 899.985801] env[65522]: DEBUG oslo_vmware.api [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 899.985801] env[65522]: value = "task-5114273" [ 899.985801] env[65522]: _type = "Task" [ 899.985801] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.991614] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cf6ea3d7-e8f0-44bc-9ef7-647f9a38b6f8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.994302] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114271, 'name': PowerOffVM_Task, 'duration_secs': 0.404615} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.003463] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.003747] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 900.004136] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 900.004415] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 900.007407] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 1e8088ef-6354-40a0-95ee-06732e7f088f] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 900.011770] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a3f8b7-4e2e-4ef7-95ba-d79cb9a954b8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.017835] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.018152] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 900.023976] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2e87310-53e4-4a4c-a95b-e1e4ef518e6d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.028247] env[65522]: DEBUG oslo_vmware.api [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114273, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.035321] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 900.036251] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-83d3ac9d-919e-4bc2-a225-c59b4f04fdb0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.041046] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Waiting for the task: (returnval){ [ 900.041046] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52704411-7265-504c-fd4e-7f2013c3de90" [ 900.041046] env[65522]: _type = "Task" [ 900.041046] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.053165] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52704411-7265-504c-fd4e-7f2013c3de90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.112468] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 900.112853] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 900.113099] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleting the datastore file [datastore2] bb20d99f-fc49-47bf-a6b7-74ea966b71eb {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.113475] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e9f28e92-8490-4765-8dc1-ba26b14ec7f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.122037] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 900.122037] env[65522]: value = "task-5114275" [ 900.122037] env[65522]: _type = "Task" [ 900.122037] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.132377] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114275, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.149681] env[65522]: DEBUG oslo_vmware.api [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114272, 'name': PowerOnVM_Task, 'duration_secs': 0.510683} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.151370] env[65522]: WARNING openstack [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 900.151751] env[65522]: WARNING openstack [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 900.159436] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 900.159703] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-84e09c08-df46-40c9-967f-26a1a6755f06 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance '5e227f6c-9f13-416f-8e6d-2f7d931619fd' progress to 100 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 900.186381] env[65522]: DEBUG nova.compute.utils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 900.187969] env[65522]: DEBUG nova.compute.manager [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 900.188185] env[65522]: DEBUG nova.network.neutron [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 900.188508] env[65522]: WARNING neutronclient.v2_0.client [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 900.188790] env[65522]: WARNING neutronclient.v2_0.client [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 900.189385] env[65522]: WARNING openstack [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 900.190705] env[65522]: WARNING openstack [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 900.256672] env[65522]: DEBUG nova.policy [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18ceaff46b314c9f9da68a8d0c6f5f2b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fa11b46d9fe144f391233e6eb9c819d7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 900.276911] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fed3bd68-87f0-40a1-8af8-03cd6ed85eb4 tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.621s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.283875] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ac70f63e-db47-46ab-90f0-7abd199fac40 tempest-ServerShowV257Test-42139899 tempest-ServerShowV257Test-42139899-project-member] Lock "26432ad6-610e-49c2-8699-f1e7e128e5b8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.563s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.304387] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55d88d68-01d3-4be5-bff1-b2ca09cdda8f tempest-MultipleCreateTestJSON-1088853482 tempest-MultipleCreateTestJSON-1088853482-project-member] Lock "179a37ed-1176-4082-8c68-1be6abd3fd9d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.124s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.305679] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0ee5627e-b659-428b-83a7-01fb84973b81 tempest-ServersTestFqdnHostnames-507733180 tempest-ServersTestFqdnHostnames-507733180-project-member] Lock "18568baa-83de-4b54-addf-dee66211c8d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.120s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.317327] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114266, 'name': CloneVM_Task, 'duration_secs': 1.713267} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.317626] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Created linked-clone VM from snapshot [ 900.318871] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baffdcdd-df0c-42df-a8a0-1e602de9e56a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.323715] env[65522]: WARNING neutronclient.v2_0.client [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 900.323715] env[65522]: WARNING openstack [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 900.323715] env[65522]: WARNING openstack [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 900.338259] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8b03c86e-ec12-4604-8bb3-fa2f0bfc6b2e tempest-ListServerFiltersTestJSON-1743219944 tempest-ListServerFiltersTestJSON-1743219944-project-member] Lock "e5d2cc50-4ee6-4e64-9b52-888968a717ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.660s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.343359] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Uploading image c1b9cca3-ccd4-41ca-bd21-b19df25ddf57 {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 900.394014] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 900.394014] env[65522]: value = "vm-994837" [ 900.394014] env[65522]: _type = "VirtualMachine" [ 900.394014] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 900.394430] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8e2c093b-b4ce-4271-8703-58e843ee4193 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.403212] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lease: (returnval){ [ 900.403212] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d9adbd-13c8-9a8c-38ee-b508c84e4fac" [ 900.403212] env[65522]: _type = "HttpNfcLease" [ 900.403212] env[65522]: } obtained for exporting VM: (result){ [ 900.403212] env[65522]: value = "vm-994837" [ 900.403212] env[65522]: _type = "VirtualMachine" [ 900.403212] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 900.403641] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the lease: (returnval){ [ 900.403641] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d9adbd-13c8-9a8c-38ee-b508c84e4fac" [ 900.403641] env[65522]: _type = "HttpNfcLease" [ 900.403641] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 900.413076] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 900.413076] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d9adbd-13c8-9a8c-38ee-b508c84e4fac" [ 900.413076] env[65522]: _type = "HttpNfcLease" [ 900.413076] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 900.509865] env[65522]: DEBUG oslo_vmware.api [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114273, 'name': ReconfigVM_Task, 'duration_secs': 0.506685} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.510330] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Reconfigured VM instance instance-0000002e to detach disk 2002 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 900.510652] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 900.511048] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9e606770-c0de-466e-a9ab-b8db8e436649 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.515146] env[65522]: DEBUG nova.compute.manager [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 900.519265] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 4238d766-2a5d-40d7-b2f6-5cb9a5354760] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 900.523998] env[65522]: DEBUG oslo_vmware.api [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 900.523998] env[65522]: value = "task-5114277" [ 900.523998] env[65522]: _type = "Task" [ 900.523998] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.537383] env[65522]: DEBUG oslo_vmware.api [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114277, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.553664] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52704411-7265-504c-fd4e-7f2013c3de90, 'name': SearchDatastore_Task, 'duration_secs': 0.020156} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.554904] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c347a9b6-de52-4b93-8b74-38a42dd602df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.563368] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Waiting for the task: (returnval){ [ 900.563368] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52627de5-b45e-744d-8649-730a0df04bb3" [ 900.563368] env[65522]: _type = "Task" [ 900.563368] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.579299] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52627de5-b45e-744d-8649-730a0df04bb3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.632768] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114275, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.698126] env[65522]: DEBUG nova.compute.manager [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 900.913344] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 900.913344] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d9adbd-13c8-9a8c-38ee-b508c84e4fac" [ 900.913344] env[65522]: _type = "HttpNfcLease" [ 900.913344] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 900.913648] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 900.913648] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d9adbd-13c8-9a8c-38ee-b508c84e4fac" [ 900.913648] env[65522]: _type = "HttpNfcLease" [ 900.913648] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 900.914441] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf5fc957-4c93-4dfb-951b-4ba54abf9dcc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.924036] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522dab6c-e699-dfc0-82f2-16b9bc7ee36d/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 900.924244] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522dab6c-e699-dfc0-82f2-16b9bc7ee36d/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 901.027019] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: ea5b8e29-ebe6-444b-bd2d-029201003ee6] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 901.043954] env[65522]: DEBUG oslo_vmware.api [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114277, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.050705] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.064170] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-234f65db-ca20-4221-80c0-7b91afde4cf2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.080126] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52627de5-b45e-744d-8649-730a0df04bb3, 'name': SearchDatastore_Task, 'duration_secs': 0.014094} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.080126] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 901.080306] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 55ab4771-9908-4640-a142-3cb40c0c6ee5/55ab4771-9908-4640-a142-3cb40c0c6ee5.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 901.080511] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d6b24b3-5bec-4a80-bd31-7cc4c8807ff7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.093476] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Waiting for the task: (returnval){ [ 901.093476] env[65522]: value = "task-5114278" [ 901.093476] env[65522]: _type = "Task" [ 901.093476] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.106791] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114278, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.135729] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114275, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.728693} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.136019] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.136209] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.136390] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.212095] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76758abf-9a1d-4159-a061-7b421d009a48 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.223850] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-264d258c-2c23-4ecc-bdea-b553ce294f30 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.267627] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e00486-206b-48f1-98a1-251b2b1fcff8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.276657] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07f4e01-ea1b-4f7c-aff8-30f5d8b779bd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.293203] env[65522]: DEBUG nova.compute.provider_tree [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 901.535332] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: da211708-a6a2-4e03-bb41-c5b93564de20] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 901.545809] env[65522]: DEBUG oslo_vmware.api [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114277, 'name': PowerOnVM_Task, 'duration_secs': 0.636844} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.545809] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 901.546235] env[65522]: DEBUG nova.compute.manager [None req-8d473b4c-d7a2-4f48-b207-60cfe38b33fc tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 901.547833] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e64455-1424-45a0-84f0-e6919a362aa4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.611674] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114278, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.708029] env[65522]: DEBUG nova.compute.manager [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 901.749659] env[65522]: DEBUG nova.virt.hardware [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 901.750740] env[65522]: DEBUG nova.virt.hardware [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 901.750740] env[65522]: DEBUG nova.virt.hardware [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 901.754154] env[65522]: DEBUG nova.virt.hardware [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 901.754444] env[65522]: DEBUG nova.virt.hardware [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 901.754721] env[65522]: DEBUG nova.virt.hardware [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 901.754992] env[65522]: DEBUG nova.virt.hardware [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.755811] env[65522]: DEBUG nova.virt.hardware [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 901.755811] env[65522]: DEBUG nova.virt.hardware [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 901.755811] env[65522]: DEBUG nova.virt.hardware [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 901.755895] env[65522]: DEBUG nova.virt.hardware [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 901.758927] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2305ad-7886-4bc6-b108-0d3bd4b7317f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.769276] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b3e79b-dafa-4407-9480-0c8b51508e67 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.817875] env[65522]: DEBUG nova.network.neutron [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Successfully created port: 54206d03-4403-4bca-8ba8-1f4e88682cb6 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 901.823395] env[65522]: ERROR nova.scheduler.client.report [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [req-76d3d36d-9de1-4342-9d78-f158d040f141] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9a0d1ac8-ed26-4657-8899-fbceab29db0d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-76d3d36d-9de1-4342-9d78-f158d040f141"}]} [ 901.848144] env[65522]: DEBUG nova.scheduler.client.report [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Refreshing inventories for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 901.860437] env[65522]: DEBUG nova.network.neutron [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Updated VIF entry in instance network info cache for port 4463f170-2656-4332-b1a3-0b0a491984da. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 901.861374] env[65522]: DEBUG nova.network.neutron [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Updating instance_info_cache with network_info: [{"id": "4463f170-2656-4332-b1a3-0b0a491984da", "address": "fa:16:3e:28:f6:07", "network": {"id": "d1672958-02a6-48dd-98df-2663748a2aa1", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-384169563-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e096164b91f8483994f1dfefbca5fdff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d88b750a-0e7d-4f16-8bd5-8e6d5743b720", "external-id": "nsx-vlan-transportzone-715", "segmentation_id": 715, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4463f170-26", "ovs_interfaceid": "4463f170-2656-4332-b1a3-0b0a491984da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 901.879044] env[65522]: DEBUG nova.scheduler.client.report [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Updating ProviderTree inventory for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 901.879044] env[65522]: DEBUG nova.compute.provider_tree [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 901.893262] env[65522]: DEBUG nova.scheduler.client.report [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Refreshing aggregate associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, aggregates: None {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 901.922176] env[65522]: DEBUG nova.scheduler.client.report [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Refreshing trait associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 902.042135] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 028c7fc5-5624-4dd5-9e2f-48191a86f765] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 902.108406] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114278, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.609157} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.108768] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 55ab4771-9908-4640-a142-3cb40c0c6ee5/55ab4771-9908-4640-a142-3cb40c0c6ee5.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 902.109064] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 902.109357] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-780f6b94-2482-41b5-987b-d74ea53d4b8a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.119049] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Waiting for the task: (returnval){ [ 902.119049] env[65522]: value = "task-5114279" [ 902.119049] env[65522]: _type = "Task" [ 902.119049] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.130633] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114279, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.186874] env[65522]: DEBUG nova.virt.hardware [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 902.187315] env[65522]: DEBUG nova.virt.hardware [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 902.187563] env[65522]: DEBUG nova.virt.hardware [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 902.187883] env[65522]: DEBUG nova.virt.hardware [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 902.188182] env[65522]: DEBUG nova.virt.hardware [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 902.188389] env[65522]: DEBUG nova.virt.hardware [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 902.188652] env[65522]: DEBUG nova.virt.hardware [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 902.188904] env[65522]: DEBUG nova.virt.hardware [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 902.189190] env[65522]: DEBUG nova.virt.hardware [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 902.189787] env[65522]: DEBUG nova.virt.hardware [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 902.189787] env[65522]: DEBUG nova.virt.hardware [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 902.190882] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf2d95f-8244-497c-be4b-fed6533644bf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.201469] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b44f5d-4e45-40fd-bb8b-d704e4430429 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.218072] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:e2:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '652babe2-668c-4364-92c2-8ceaba23c0ac', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 902.226526] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 902.231122] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 902.231864] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca713a96-be69-4ca4-966f-3e75bd3155f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.259513] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 902.259513] env[65522]: value = "task-5114280" [ 902.259513] env[65522]: _type = "Task" [ 902.259513] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.269847] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114280, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.367030] env[65522]: DEBUG oslo_concurrency.lockutils [req-65db4386-ed0f-4b76-986c-f0bfdf468b9c req-39f12cfb-b624-4ae4-ad09-9854128a4898 service nova] Releasing lock "refresh_cache-55ab4771-9908-4640-a142-3cb40c0c6ee5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 902.410040] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b9b7a1-238d-49d9-843b-cbbecb48b71f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.426047] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2967d3-e8ba-41bf-8581-37bb0acf0fb9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.471221] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9caf9416-ffd1-40b5-9546-8d7e6498c34e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.480456] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e35f2a-ec89-4f2e-883b-e94f96770aa2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.500675] env[65522]: DEBUG nova.compute.provider_tree [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.548229] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 54ab9b82-ad31-4e3b-a3f0-5b4c6d35f824] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 902.630613] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114279, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076441} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.631905] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 902.632977] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-810f4502-5a25-4b95-a13f-ac0993655b3b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.665072] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 55ab4771-9908-4640-a142-3cb40c0c6ee5/55ab4771-9908-4640-a142-3cb40c0c6ee5.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.665072] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6131f70d-f709-4526-8f76-870100822340 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.692227] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Waiting for the task: (returnval){ [ 902.692227] env[65522]: value = "task-5114281" [ 902.692227] env[65522]: _type = "Task" [ 902.692227] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.704551] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114281, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.773671] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114280, 'name': CreateVM_Task, 'duration_secs': 0.400694} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.774288] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 902.774437] env[65522]: WARNING neutronclient.v2_0.client [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 902.774846] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.774994] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 902.775364] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 902.775652] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-389d60ac-9541-42f6-85fa-e629682623e1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.783774] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 902.783774] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524e3ea9-6dbc-59b4-970b-b689a2ced388" [ 902.783774] env[65522]: _type = "Task" [ 902.783774] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.802250] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524e3ea9-6dbc-59b4-970b-b689a2ced388, 'name': SearchDatastore_Task, 'duration_secs': 0.011909} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.802250] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 902.803780] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.803780] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.803780] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 902.803780] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.804390] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-027bd464-0edb-4ce1-bb7b-32e8d4555994 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.817112] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.817369] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.818372] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4bdca49-c726-4f34-a179-b2b9c356f237 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.831128] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 902.831128] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ea0625-6efb-63e0-671c-c4822074f450" [ 902.831128] env[65522]: _type = "Task" [ 902.831128] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.846681] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ea0625-6efb-63e0-671c-c4822074f450, 'name': SearchDatastore_Task, 'duration_secs': 0.015576} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.848541] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d574f94d-c2d4-4ae3-a816-599fbe5ca6d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.859135] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 902.859135] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523f0607-e565-e962-e15c-f9e757f00c85" [ 902.859135] env[65522]: _type = "Task" [ 902.859135] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.872675] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523f0607-e565-e962-e15c-f9e757f00c85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.004843] env[65522]: DEBUG nova.scheduler.client.report [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 903.048981] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 428873b8-1e58-4924-86f3-8feca8f53342] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 903.203387] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114281, 'name': ReconfigVM_Task, 'duration_secs': 0.356963} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.203690] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 55ab4771-9908-4640-a142-3cb40c0c6ee5/55ab4771-9908-4640-a142-3cb40c0c6ee5.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 903.204377] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-74399637-4813-49dd-bdb6-6e756461dfde {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.213600] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Waiting for the task: (returnval){ [ 903.213600] env[65522]: value = "task-5114282" [ 903.213600] env[65522]: _type = "Task" [ 903.213600] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.223674] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114282, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.373162] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523f0607-e565-e962-e15c-f9e757f00c85, 'name': SearchDatastore_Task, 'duration_secs': 0.018599} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.373507] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 903.373828] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] bb20d99f-fc49-47bf-a6b7-74ea966b71eb/bb20d99f-fc49-47bf-a6b7-74ea966b71eb.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.374173] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21eb5596-3beb-4fed-87e8-8e22b9bbe2b2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.383815] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 903.383815] env[65522]: value = "task-5114283" [ 903.383815] env[65522]: _type = "Task" [ 903.383815] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.395660] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114283, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.513480] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.818s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.514120] env[65522]: DEBUG nova.compute.manager [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 903.517386] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.913s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 903.517707] env[65522]: DEBUG nova.objects.instance [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lazy-loading 'resources' on Instance uuid 495fa98f-a9db-4214-87cc-d29209d3cb62 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.651534] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 361e54c7-4309-48b3-ad3e-b2ebc783741c] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 903.730698] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114282, 'name': Rename_Task, 'duration_secs': 0.167043} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.731114] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 903.731478] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b0e29d50-b7c5-447a-b7a2-ecfba545f6af {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.743341] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Waiting for the task: (returnval){ [ 903.743341] env[65522]: value = "task-5114284" [ 903.743341] env[65522]: _type = "Task" [ 903.743341] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.756708] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114284, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.899642] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114283, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.152894] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 20855931-a02d-4b53-b998-3d89d4ff5d1a] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 904.156071] env[65522]: DEBUG nova.compute.utils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 904.160971] env[65522]: DEBUG nova.compute.manager [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 904.161221] env[65522]: DEBUG nova.network.neutron [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 904.161652] env[65522]: WARNING neutronclient.v2_0.client [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.162330] env[65522]: WARNING neutronclient.v2_0.client [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.163089] env[65522]: WARNING openstack [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.163607] env[65522]: WARNING openstack [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.260229] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114284, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.400911] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114283, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.648177} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.401645] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] bb20d99f-fc49-47bf-a6b7-74ea966b71eb/bb20d99f-fc49-47bf-a6b7-74ea966b71eb.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 904.402569] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.403027] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-58806369-5bdb-4ec2-9ed2-ab8ad6acb8f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.412464] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 904.412464] env[65522]: value = "task-5114285" [ 904.412464] env[65522]: _type = "Task" [ 904.412464] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.426998] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114285, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.428503] env[65522]: DEBUG nova.network.neutron [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Successfully updated port: 54206d03-4403-4bca-8ba8-1f4e88682cb6 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 904.474547] env[65522]: DEBUG nova.policy [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e74c132f591d4c6fbe65a75c1e1e4a24', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ecccb656b0d4c96b40b200cdcddbad5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 904.598467] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81d79ec-b4c5-4f5c-a0f2-c7e606bc286e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.607903] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7543032b-ca52-485c-8571-8d5babf38d02 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.646514] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcddc768-27b6-492f-8630-c67914e695af {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.656267] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9eaf8d5-91c4-4293-9500-c6ba9d18e3a4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.661088] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: c0c99008-1907-4d3a-b4f6-ee6e916e2c67] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 904.663508] env[65522]: DEBUG nova.compute.manager [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 904.675943] env[65522]: DEBUG nova.compute.provider_tree [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.756726] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114284, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.927534] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114285, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080591} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.927844] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.928790] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9861ef-aa4b-4e6c-b451-b6faa7948c13 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.938758] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-9c39b722-56ff-44fa-8f66-3e3432645a68" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.938758] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-9c39b722-56ff-44fa-8f66-3e3432645a68" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 904.938758] env[65522]: DEBUG nova.network.neutron [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 904.961648] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] bb20d99f-fc49-47bf-a6b7-74ea966b71eb/bb20d99f-fc49-47bf-a6b7-74ea966b71eb.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.962773] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f5f3967-36a7-4127-9959-7d276429fa8f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.989835] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 904.989835] env[65522]: value = "task-5114286" [ 904.989835] env[65522]: _type = "Task" [ 904.989835] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.004957] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114286, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.166410] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 6e10e846-cdb7-458e-b511-18ab1742228e] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 905.169344] env[65522]: DEBUG nova.network.neutron [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Successfully created port: f8145f30-4e64-4a22-89a7-ffa393891ffb {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 905.181023] env[65522]: DEBUG nova.scheduler.client.report [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 905.263773] env[65522]: DEBUG oslo_vmware.api [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114284, 'name': PowerOnVM_Task, 'duration_secs': 1.145556} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.264737] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 905.264957] env[65522]: INFO nova.compute.manager [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Took 9.44 seconds to spawn the instance on the hypervisor. [ 905.265173] env[65522]: DEBUG nova.compute.manager [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 905.266011] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611ed0bf-4892-4ab8-9bd5-ea654c650911 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.435404] env[65522]: WARNING openstack [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 905.435788] env[65522]: WARNING openstack [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 905.501237] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114286, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.678392] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: f44db343-7e17-4780-9a67-245992ac1a63] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 905.683373] env[65522]: DEBUG nova.compute.manager [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 905.688795] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.171s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 905.691797] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.086s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 905.692121] env[65522]: DEBUG nova.objects.instance [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lazy-loading 'resources' on Instance uuid 308330f8-1219-4c02-a129-22b29f0e33b2 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 905.718197] env[65522]: DEBUG nova.virt.hardware [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 905.718883] env[65522]: DEBUG nova.virt.hardware [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 905.719138] env[65522]: DEBUG nova.virt.hardware [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 905.719436] env[65522]: DEBUG nova.virt.hardware [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 905.719713] env[65522]: DEBUG nova.virt.hardware [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 905.719975] env[65522]: DEBUG nova.virt.hardware [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 905.720333] env[65522]: DEBUG nova.virt.hardware [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.720589] env[65522]: DEBUG nova.virt.hardware [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 905.720873] env[65522]: DEBUG nova.virt.hardware [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 905.721131] env[65522]: DEBUG nova.virt.hardware [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 905.721451] env[65522]: DEBUG nova.virt.hardware [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 905.723408] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d27f61-d390-4439-9b29-b1b7677a8784 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.727420] env[65522]: INFO nova.scheduler.client.report [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleted allocations for instance 495fa98f-a9db-4214-87cc-d29209d3cb62 [ 905.736839] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a03509-c5e2-4947-9ca2-cf50a938d5d7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.790639] env[65522]: INFO nova.compute.manager [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Took 37.33 seconds to build instance. [ 906.002084] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114286, 'name': ReconfigVM_Task, 'duration_secs': 0.870058} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.002660] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Reconfigured VM instance instance-00000029 to attach disk [datastore1] bb20d99f-fc49-47bf-a6b7-74ea966b71eb/bb20d99f-fc49-47bf-a6b7-74ea966b71eb.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.003071] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ae370362-01b1-4479-aa82-9df4e57c0afb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.012636] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 906.012636] env[65522]: value = "task-5114287" [ 906.012636] env[65522]: _type = "Task" [ 906.012636] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.025520] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114287, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.175808] env[65522]: DEBUG nova.network.neutron [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 906.184448] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: dbf29bd1-780c-4756-a203-4fad05a9d3b9] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 906.240572] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6a76d39f-6f07-4d11-98cf-1eca2a791118 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "495fa98f-a9db-4214-87cc-d29209d3cb62" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.429s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.294833] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0291cb58-7934-4a5f-82d4-b1218a2b7ed8 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Lock "55ab4771-9908-4640-a142-3cb40c0c6ee5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.860s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.523161] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114287, 'name': Rename_Task, 'duration_secs': 0.35637} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.526082] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 906.526593] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-92e3e1e9-40e9-48c5-a537-7376a44b084e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.534128] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 906.534128] env[65522]: value = "task-5114288" [ 906.534128] env[65522]: _type = "Task" [ 906.534128] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.548482] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114288, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.607103] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1282e10-de03-415a-ae84-d0edcafff34c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.616220] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a040ab1a-fb56-444a-b1c3-62b7a6423b68 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.653997] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637794e6-6b7c-457d-9640-2b80d2434166 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.662657] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3d89cc-ea0d-4f35-8d44-a91a41b13211 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.677822] env[65522]: DEBUG nova.compute.provider_tree [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.688494] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: f5428ab0-a7ec-4ed0-8a03-dbcd584f22f7] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 907.034382] env[65522]: DEBUG nova.network.neutron [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Successfully updated port: f8145f30-4e64-4a22-89a7-ffa393891ffb {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 907.052403] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114288, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.181236] env[65522]: DEBUG nova.scheduler.client.report [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 907.192846] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: cd82f2fd-d657-4ece-8e90-9a47072f5546] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 907.259583] env[65522]: WARNING openstack [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 907.259989] env[65522]: WARNING openstack [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 907.545084] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.545312] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 907.545491] env[65522]: DEBUG nova.network.neutron [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 907.551869] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114288, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.687232] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.995s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 907.931048] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 12.055s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 907.931048] env[65522]: DEBUG nova.objects.instance [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 907.931048] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 97dc8cac-8241-4912-a3ed-689439ef7ff8] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 907.989314] env[65522]: INFO nova.scheduler.client.report [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Deleted allocations for instance 308330f8-1219-4c02-a129-22b29f0e33b2 [ 908.376898] env[65522]: WARNING openstack [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 908.376898] env[65522]: WARNING openstack [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 908.400014] env[65522]: DEBUG oslo_vmware.api [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114288, 'name': PowerOnVM_Task, 'duration_secs': 1.19433} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.400395] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 908.400603] env[65522]: DEBUG nova.compute.manager [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 908.401483] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3339b9ba-8249-429e-ba45-caa84556349e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.438405] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: a8f9f223-023a-48fc-9186-a360a78e27a9] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 908.498713] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4657946f-1f69-4e40-8c5e-7b5d03d9abcf tempest-ImagesOneServerNegativeTestJSON-186561115 tempest-ImagesOneServerNegativeTestJSON-186561115-project-member] Lock "308330f8-1219-4c02-a129-22b29f0e33b2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.692s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.794089] env[65522]: WARNING neutronclient.v2_0.client [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 908.795159] env[65522]: WARNING openstack [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 908.795159] env[65522]: WARNING openstack [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 908.831745] env[65522]: DEBUG nova.network.neutron [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 908.927806] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 908.941260] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bae9050d-0bca-495a-aca9-ac9cb3f3085e tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.248s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.941839] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.636s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 908.942203] env[65522]: DEBUG nova.objects.instance [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Lazy-loading 'resources' on Instance uuid 9da312a0-c6eb-4903-a7d8-e178aa54d2ea {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.943371] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 0d52b207-ac69-48ce-b3af-7a83a499886e] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 909.076851] env[65522]: DEBUG nova.compute.manager [req-20fb93ac-0122-4288-875a-395f7675d2e1 req-09940d06-57f4-4ecf-b285-926f73dea85b service nova] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Received event network-vif-plugged-54206d03-4403-4bca-8ba8-1f4e88682cb6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 909.077119] env[65522]: DEBUG oslo_concurrency.lockutils [req-20fb93ac-0122-4288-875a-395f7675d2e1 req-09940d06-57f4-4ecf-b285-926f73dea85b service nova] Acquiring lock "9c39b722-56ff-44fa-8f66-3e3432645a68-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 909.077348] env[65522]: DEBUG oslo_concurrency.lockutils [req-20fb93ac-0122-4288-875a-395f7675d2e1 req-09940d06-57f4-4ecf-b285-926f73dea85b service nova] Lock "9c39b722-56ff-44fa-8f66-3e3432645a68-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 909.077496] env[65522]: DEBUG oslo_concurrency.lockutils [req-20fb93ac-0122-4288-875a-395f7675d2e1 req-09940d06-57f4-4ecf-b285-926f73dea85b service nova] Lock "9c39b722-56ff-44fa-8f66-3e3432645a68-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 909.077659] env[65522]: DEBUG nova.compute.manager [req-20fb93ac-0122-4288-875a-395f7675d2e1 req-09940d06-57f4-4ecf-b285-926f73dea85b service nova] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] No waiting events found dispatching network-vif-plugged-54206d03-4403-4bca-8ba8-1f4e88682cb6 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 909.077819] env[65522]: WARNING nova.compute.manager [req-20fb93ac-0122-4288-875a-395f7675d2e1 req-09940d06-57f4-4ecf-b285-926f73dea85b service nova] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Received unexpected event network-vif-plugged-54206d03-4403-4bca-8ba8-1f4e88682cb6 for instance with vm_state building and task_state spawning. [ 909.382604] env[65522]: WARNING openstack [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 909.382604] env[65522]: WARNING openstack [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 909.643868] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: fdf56fa6-ad18-499e-a171-6242abb293bf] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 909.652165] env[65522]: WARNING neutronclient.v2_0.client [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 909.982177] env[65522]: DEBUG nova.network.neutron [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Updating instance_info_cache with network_info: [{"id": "54206d03-4403-4bca-8ba8-1f4e88682cb6", "address": "fa:16:3e:72:7b:af", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54206d03-44", "ovs_interfaceid": "54206d03-4403-4bca-8ba8-1f4e88682cb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 910.041144] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cab9432-6fed-4819-82bd-778438f9c14e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.051168] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994e81ad-3700-4c54-b057-3a71b094da9c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.091625] env[65522]: WARNING neutronclient.v2_0.client [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 910.091995] env[65522]: WARNING neutronclient.v2_0.client [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 910.096873] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c370817-d59f-448b-9024-7e6820e385d6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.103871] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7173ef76-ced4-45e7-930f-2b6773556111 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.118682] env[65522]: DEBUG nova.compute.provider_tree [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.148310] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: ad074365-1443-481d-990f-479c523219fa] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 910.364828] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "4aca4e87-b923-49e4-88d8-7ff51bcbe993" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.365148] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "4aca4e87-b923-49e4-88d8-7ff51bcbe993" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.489183] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-9c39b722-56ff-44fa-8f66-3e3432645a68" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 910.489810] env[65522]: DEBUG nova.compute.manager [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Instance network_info: |[{"id": "54206d03-4403-4bca-8ba8-1f4e88682cb6", "address": "fa:16:3e:72:7b:af", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54206d03-44", "ovs_interfaceid": "54206d03-4403-4bca-8ba8-1f4e88682cb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 910.490436] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:7b:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4307c18-b235-43cd-bcd5-e226012d8ee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '54206d03-4403-4bca-8ba8-1f4e88682cb6', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 910.503086] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 910.504534] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 910.504534] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ea648ec-0886-43b1-9429-ee0f563f4a67 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.537141] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 910.537141] env[65522]: value = "task-5114289" [ 910.537141] env[65522]: _type = "Task" [ 910.537141] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.550304] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114289, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.588036] env[65522]: WARNING neutronclient.v2_0.client [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 910.588036] env[65522]: WARNING openstack [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 910.588036] env[65522]: WARNING openstack [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 910.622551] env[65522]: DEBUG nova.scheduler.client.report [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 910.637380] env[65522]: DEBUG nova.network.neutron [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Port 3f79042c-23dd-4fc0-b3cf-1e2884edabc1 binding to destination host cpu-1 is already ACTIVE {{(pid=65522) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 910.637380] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.637380] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 910.637712] env[65522]: DEBUG nova.network.neutron [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 910.653369] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 6c8165f5-6769-4954-b9ac-de8551954377] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 910.865463] env[65522]: DEBUG nova.network.neutron [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating instance_info_cache with network_info: [{"id": "f8145f30-4e64-4a22-89a7-ffa393891ffb", "address": "fa:16:3e:42:70:f5", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8145f30-4e", "ovs_interfaceid": "f8145f30-4e64-4a22-89a7-ffa393891ffb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 910.869220] env[65522]: DEBUG nova.compute.manager [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 910.881859] env[65522]: DEBUG nova.compute.manager [req-42a66a01-ca9f-400a-862a-25e88970d892 req-85937f24-6272-4a17-bd92-0106337b85ef service nova] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Received event network-vif-plugged-f8145f30-4e64-4a22-89a7-ffa393891ffb {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 910.882580] env[65522]: DEBUG oslo_concurrency.lockutils [req-42a66a01-ca9f-400a-862a-25e88970d892 req-85937f24-6272-4a17-bd92-0106337b85ef service nova] Acquiring lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.882580] env[65522]: DEBUG oslo_concurrency.lockutils [req-42a66a01-ca9f-400a-862a-25e88970d892 req-85937f24-6272-4a17-bd92-0106337b85ef service nova] Lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.882580] env[65522]: DEBUG oslo_concurrency.lockutils [req-42a66a01-ca9f-400a-862a-25e88970d892 req-85937f24-6272-4a17-bd92-0106337b85ef service nova] Lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.882772] env[65522]: DEBUG nova.compute.manager [req-42a66a01-ca9f-400a-862a-25e88970d892 req-85937f24-6272-4a17-bd92-0106337b85ef service nova] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] No waiting events found dispatching network-vif-plugged-f8145f30-4e64-4a22-89a7-ffa393891ffb {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 910.882933] env[65522]: WARNING nova.compute.manager [req-42a66a01-ca9f-400a-862a-25e88970d892 req-85937f24-6272-4a17-bd92-0106337b85ef service nova] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Received unexpected event network-vif-plugged-f8145f30-4e64-4a22-89a7-ffa393891ffb for instance with vm_state building and task_state spawning. [ 911.273317] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.331s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 911.280228] env[65522]: WARNING neutronclient.v2_0.client [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 911.280228] env[65522]: WARNING openstack [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 911.281337] env[65522]: WARNING openstack [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 911.290973] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 2b1b2fa1-cd4e-4359-bbe0-d7ca16d0cf20] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 911.300622] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.249s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 911.304114] env[65522]: INFO nova.compute.claims [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 911.321958] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114289, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.341368] env[65522]: INFO nova.scheduler.client.report [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Deleted allocations for instance 9da312a0-c6eb-4903-a7d8-e178aa54d2ea [ 911.384640] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 911.384640] env[65522]: DEBUG nova.compute.manager [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Instance network_info: |[{"id": "f8145f30-4e64-4a22-89a7-ffa393891ffb", "address": "fa:16:3e:42:70:f5", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8145f30-4e", "ovs_interfaceid": "f8145f30-4e64-4a22-89a7-ffa393891ffb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 911.390022] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:70:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6fb0104-186b-4288-b87e-634893f46f01', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f8145f30-4e64-4a22-89a7-ffa393891ffb', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 911.397848] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 911.403379] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 911.404040] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-75c5598c-7d71-40b4-af24-78c664db69ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.421239] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 911.429046] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 911.429046] env[65522]: value = "task-5114290" [ 911.429046] env[65522]: _type = "Task" [ 911.429046] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.438805] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114290, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.130534] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 3f749f68-879f-44f3-8477-7ecea947078c] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 912.143961] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3681b6ec-fad4-4702-89c4-f1b10eae3ebe tempest-InstanceActionsV221TestJSON-902505420 tempest-InstanceActionsV221TestJSON-902505420-project-member] Lock "9da312a0-c6eb-4903-a7d8-e178aa54d2ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.949s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.174348] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114290, 'name': CreateVM_Task, 'duration_secs': 0.66021} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.174839] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114289, 'name': CreateVM_Task, 'duration_secs': 0.817478} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.174839] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 912.175060] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 912.177075] env[65522]: WARNING neutronclient.v2_0.client [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 912.177495] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.177686] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 912.180994] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 912.181562] env[65522]: WARNING neutronclient.v2_0.client [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 912.181879] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.182124] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e92a343-7830-4f32-a556-242a1472052f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.189503] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 912.189503] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f55ded-c07f-f8ff-2e2b-57c0c6c92dc2" [ 912.189503] env[65522]: _type = "Task" [ 912.189503] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.202777] env[65522]: WARNING openstack [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.202777] env[65522]: WARNING openstack [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.219459] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f55ded-c07f-f8ff-2e2b-57c0c6c92dc2, 'name': SearchDatastore_Task, 'duration_secs': 0.016042} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.219790] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 912.220785] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 912.224084] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.224084] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 912.224084] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 912.224084] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 912.224084] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 912.224247] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fccb0180-d111-45ae-b7aa-c7ee888a6814 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.227257] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f805e4e-a1e4-42fd-b3d7-f37db149a403 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.236499] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 912.236499] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f19b69-b1e4-179a-1f35-fc591387de94" [ 912.236499] env[65522]: _type = "Task" [ 912.236499] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.252899] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f19b69-b1e4-179a-1f35-fc591387de94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.255600] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 912.255712] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 912.258526] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1adda26-9cfa-4eb1-8c73-38d7a7ced072 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.263863] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 912.263863] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524596e3-3c1b-ecc9-8946-c8f7ef4cdd8d" [ 912.263863] env[65522]: _type = "Task" [ 912.263863] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.273822] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524596e3-3c1b-ecc9-8946-c8f7ef4cdd8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.382151] env[65522]: WARNING neutronclient.v2_0.client [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 912.382844] env[65522]: WARNING openstack [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.383207] env[65522]: WARNING openstack [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.477345] env[65522]: DEBUG nova.network.neutron [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance_info_cache with network_info: [{"id": "3f79042c-23dd-4fc0-b3cf-1e2884edabc1", "address": "fa:16:3e:f7:f5:9e", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.131", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f79042c-23", "ovs_interfaceid": "3f79042c-23dd-4fc0-b3cf-1e2884edabc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 912.824193] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: a09c12be-0aaa-4934-a1b3-5af29feba8a6] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 912.855707] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524596e3-3c1b-ecc9-8946-c8f7ef4cdd8d, 'name': SearchDatastore_Task, 'duration_secs': 0.017523} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.856090] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f19b69-b1e4-179a-1f35-fc591387de94, 'name': SearchDatastore_Task, 'duration_secs': 0.015102} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.857503] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 912.857840] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 912.858130] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.859499] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9aca97aa-82b2-46a8-89e3-1d4f70ff07d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.871603] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 912.871603] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5222f9db-3c0a-7345-4e1d-99918f50fc56" [ 912.871603] env[65522]: _type = "Task" [ 912.871603] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.887808] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5222f9db-3c0a-7345-4e1d-99918f50fc56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.982916] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Releasing lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 913.141801] env[65522]: DEBUG nova.compute.manager [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Received event network-changed-54206d03-4403-4bca-8ba8-1f4e88682cb6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 913.142251] env[65522]: DEBUG nova.compute.manager [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Refreshing instance network info cache due to event network-changed-54206d03-4403-4bca-8ba8-1f4e88682cb6. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 913.142251] env[65522]: DEBUG oslo_concurrency.lockutils [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] Acquiring lock "refresh_cache-9c39b722-56ff-44fa-8f66-3e3432645a68" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.142376] env[65522]: DEBUG oslo_concurrency.lockutils [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] Acquired lock "refresh_cache-9c39b722-56ff-44fa-8f66-3e3432645a68" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 913.142572] env[65522]: DEBUG nova.network.neutron [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Refreshing network info cache for port 54206d03-4403-4bca-8ba8-1f4e88682cb6 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 913.331858] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: bc40780e-43e2-4058-ab6f-d0d65e839128] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 913.477152] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Acquiring lock "55ab4771-9908-4640-a142-3cb40c0c6ee5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 913.477152] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Lock "55ab4771-9908-4640-a142-3cb40c0c6ee5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 913.477152] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Acquiring lock "55ab4771-9908-4640-a142-3cb40c0c6ee5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 913.477152] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Lock "55ab4771-9908-4640-a142-3cb40c0c6ee5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 913.477468] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Lock "55ab4771-9908-4640-a142-3cb40c0c6ee5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 913.482407] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5222f9db-3c0a-7345-4e1d-99918f50fc56, 'name': SearchDatastore_Task, 'duration_secs': 0.012665} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.482407] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 913.483781] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] a2a5fb2d-62e0-4809-a01f-f1df66dca58f/a2a5fb2d-62e0-4809-a01f-f1df66dca58f.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 913.485036] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c725bb6-2f3d-48d8-90de-22018ec202a0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.489784] env[65522]: INFO nova.compute.manager [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Terminating instance [ 913.492111] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 913.492424] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.492732] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1cf005b-8da9-4031-92b3-ba048688a5ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.499829] env[65522]: DEBUG nova.compute.manager [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65522) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:924}} [ 913.500272] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 913.508033] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5970bbaa-0b82-4298-b50a-d2a2f0f59d1d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.514993] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a64ef41-08fa-46e4-9d5e-2f0bf7619f37 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.525718] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.525948] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 913.526724] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 913.526724] env[65522]: value = "task-5114291" [ 913.526724] env[65522]: _type = "Task" [ 913.526724] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.527446] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a29211ba-9c8d-4a26-bb46-6a11b07d419f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.565407] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41370abf-6294-43f0-8b80-fc202c359758 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.573317] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 913.573317] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522b4089-f8d6-9433-9725-c6276bc8e139" [ 913.573317] env[65522]: _type = "Task" [ 913.573317] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.574022] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114291, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.585219] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1da1bf9-aba7-4efb-988e-7aeb8ca931d0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.591369] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522b4089-f8d6-9433-9725-c6276bc8e139, 'name': SearchDatastore_Task, 'duration_secs': 0.01106} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.591369] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54f16475-9a4a-4664-acf3-9c03f997dcfd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.601784] env[65522]: DEBUG nova.compute.provider_tree [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.608111] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 913.608111] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e7d6e7-b46a-63c3-ba42-699876aa3ca2" [ 913.608111] env[65522]: _type = "Task" [ 913.608111] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.622419] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e7d6e7-b46a-63c3-ba42-699876aa3ca2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.646138] env[65522]: WARNING neutronclient.v2_0.client [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.646589] env[65522]: WARNING openstack [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.647361] env[65522]: WARNING openstack [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.704229] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Acquiring lock "2a12a0a3-913e-4ade-85ef-cca7ba508ac9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 913.704479] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Lock "2a12a0a3-913e-4ade-85ef-cca7ba508ac9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 913.760219] env[65522]: DEBUG nova.compute.manager [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Received event network-changed-f8145f30-4e64-4a22-89a7-ffa393891ffb {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 913.760219] env[65522]: DEBUG nova.compute.manager [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Refreshing instance network info cache due to event network-changed-f8145f30-4e64-4a22-89a7-ffa393891ffb. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 913.760521] env[65522]: DEBUG oslo_concurrency.lockutils [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Acquiring lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.760616] env[65522]: DEBUG oslo_concurrency.lockutils [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Acquired lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 913.760834] env[65522]: DEBUG nova.network.neutron [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Refreshing network info cache for port f8145f30-4e64-4a22-89a7-ffa393891ffb {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 913.800396] env[65522]: WARNING openstack [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.800590] env[65522]: WARNING openstack [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.908011] env[65522]: WARNING neutronclient.v2_0.client [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.908798] env[65522]: WARNING openstack [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.909192] env[65522]: WARNING openstack [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.979943] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 6abf58ca-ad6b-4fe8-844a-ca33e62a844d] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 914.012344] env[65522]: DEBUG nova.compute.manager [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 914.012344] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 914.013659] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39fc8946-9d11-44bb-8f3f-aaf99ce5fba3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.018964] env[65522]: DEBUG nova.network.neutron [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Updated VIF entry in instance network info cache for port 54206d03-4403-4bca-8ba8-1f4e88682cb6. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 914.019375] env[65522]: DEBUG nova.network.neutron [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Updating instance_info_cache with network_info: [{"id": "54206d03-4403-4bca-8ba8-1f4e88682cb6", "address": "fa:16:3e:72:7b:af", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54206d03-44", "ovs_interfaceid": "54206d03-4403-4bca-8ba8-1f4e88682cb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 914.029251] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 914.029617] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b08cc9e9-9624-4025-a7e5-0d494e4a397d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.043881] env[65522]: DEBUG oslo_vmware.api [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Waiting for the task: (returnval){ [ 914.043881] env[65522]: value = "task-5114292" [ 914.043881] env[65522]: _type = "Task" [ 914.043881] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.048782] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114291, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.068763] env[65522]: DEBUG oslo_vmware.api [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114292, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.105871] env[65522]: DEBUG nova.scheduler.client.report [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 914.122216] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e7d6e7-b46a-63c3-ba42-699876aa3ca2, 'name': SearchDatastore_Task, 'duration_secs': 0.015683} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.122216] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 914.122216] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 9c39b722-56ff-44fa-8f66-3e3432645a68/9c39b722-56ff-44fa-8f66-3e3432645a68.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 914.122458] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1e78e8e3-0d5d-41be-80b3-0ecf0677116b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.133935] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 914.133935] env[65522]: value = "task-5114293" [ 914.133935] env[65522]: _type = "Task" [ 914.133935] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.145394] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114293, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.207967] env[65522]: DEBUG nova.compute.manager [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 914.269125] env[65522]: WARNING neutronclient.v2_0.client [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 914.269787] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.270350] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.281600] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Acquiring lock "21715aa5-24d4-423b-92a0-be7cc1f60877" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 914.282058] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Lock "21715aa5-24d4-423b-92a0-be7cc1f60877" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 914.483043] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 1d3a9ada-df46-46af-a7f1-a84c251af4ac] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 914.524020] env[65522]: DEBUG oslo_concurrency.lockutils [req-8bd87b90-28cf-4152-9063-1213ef39c99d req-fdae066e-a004-4120-943b-debd13468bcd service nova] Releasing lock "refresh_cache-9c39b722-56ff-44fa-8f66-3e3432645a68" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 914.546098] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114291, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.559917] env[65522]: DEBUG oslo_vmware.api [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114292, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.616294] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.316s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 914.616983] env[65522]: DEBUG nova.compute.manager [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 914.620482] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 5.693s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 914.620678] env[65522]: DEBUG nova.objects.instance [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 914.647284] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114293, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.740285] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 914.787419] env[65522]: DEBUG nova.compute.manager [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 914.830298] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.830853] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.949715] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522dab6c-e699-dfc0-82f2-16b9bc7ee36d/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 914.950511] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab86164-0e74-4602-aca9-26e71061117b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.955984] env[65522]: WARNING neutronclient.v2_0.client [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 914.956764] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.957187] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.970328] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522dab6c-e699-dfc0-82f2-16b9bc7ee36d/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 914.970603] env[65522]: ERROR oslo_vmware.rw_handles [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522dab6c-e699-dfc0-82f2-16b9bc7ee36d/disk-0.vmdk due to incomplete transfer. [ 914.972274] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ac5faaaa-ec53-49a5-8255-909cc274b32c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.993714] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: b2d70982-54e7-459c-a0d8-48bf4b6e4345] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 914.993856] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522dab6c-e699-dfc0-82f2-16b9bc7ee36d/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 914.994416] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Uploaded image c1b9cca3-ccd4-41ca-bd21-b19df25ddf57 to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 914.996805] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 914.997968] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7e0139bd-2930-4cd7-b8d7-3bdc22ceca1a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.008031] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 915.008031] env[65522]: value = "task-5114294" [ 915.008031] env[65522]: _type = "Task" [ 915.008031] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.023671] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114294, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.043986] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114291, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.183296} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.046081] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] a2a5fb2d-62e0-4809-a01f-f1df66dca58f/a2a5fb2d-62e0-4809-a01f-f1df66dca58f.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 915.046081] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.046081] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-403707b0-8c77-4234-bba8-6f78d1b71043 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.056625] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 915.056625] env[65522]: value = "task-5114295" [ 915.056625] env[65522]: _type = "Task" [ 915.056625] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.066686] env[65522]: DEBUG oslo_vmware.api [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114292, 'name': PowerOffVM_Task, 'duration_secs': 0.818325} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.074205] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 915.074205] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 915.074205] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114295, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.074205] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af32fea3-ca64-411c-9b6c-abad5393dfcf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.125788] env[65522]: DEBUG nova.compute.utils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 915.129058] env[65522]: DEBUG nova.compute.manager [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 915.129058] env[65522]: DEBUG nova.network.neutron [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 915.130433] env[65522]: WARNING neutronclient.v2_0.client [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 915.130433] env[65522]: WARNING neutronclient.v2_0.client [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 915.132677] env[65522]: WARNING openstack [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 915.132875] env[65522]: WARNING openstack [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 915.160744] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114293, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.165906] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 915.166194] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 915.166405] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Deleting the datastore file [datastore1] 55ab4771-9908-4640-a142-3cb40c0c6ee5 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 915.166999] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dec8cc29-1e79-4438-b3da-5b7119a56746 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.173835] env[65522]: DEBUG nova.network.neutron [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updated VIF entry in instance network info cache for port f8145f30-4e64-4a22-89a7-ffa393891ffb. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 915.174522] env[65522]: DEBUG nova.network.neutron [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating instance_info_cache with network_info: [{"id": "f8145f30-4e64-4a22-89a7-ffa393891ffb", "address": "fa:16:3e:42:70:f5", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8145f30-4e", "ovs_interfaceid": "f8145f30-4e64-4a22-89a7-ffa393891ffb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 915.189967] env[65522]: DEBUG oslo_vmware.api [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Waiting for the task: (returnval){ [ 915.189967] env[65522]: value = "task-5114297" [ 915.189967] env[65522]: _type = "Task" [ 915.189967] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.213190] env[65522]: DEBUG oslo_vmware.api [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114297, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.265989] env[65522]: DEBUG nova.policy [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '48bde7741f2d4bfabdb803d10cf79599', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a80f979f3dc0477e9462b47f7aa87f14', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 915.537050] env[65522]: DEBUG oslo_concurrency.lockutils [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "75c8848b-aea9-43f1-8697-9224050d1fef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.537050] env[65522]: DEBUG oslo_concurrency.lockutils [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "75c8848b-aea9-43f1-8697-9224050d1fef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 915.537050] env[65522]: DEBUG oslo_concurrency.lockutils [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "75c8848b-aea9-43f1-8697-9224050d1fef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.537357] env[65522]: DEBUG oslo_concurrency.lockutils [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "75c8848b-aea9-43f1-8697-9224050d1fef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 915.537621] env[65522]: DEBUG oslo_concurrency.lockutils [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "75c8848b-aea9-43f1-8697-9224050d1fef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 915.539520] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: fbb035b5-4562-44e3-9b95-8cf85299bb43] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 915.542196] env[65522]: INFO nova.compute.manager [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Terminating instance [ 915.556496] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114294, 'name': Destroy_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.558219] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.572692] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114295, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.132488} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.572974] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 915.576020] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab18aef-b5d2-4e34-b46d-b81906a7d5ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.608480] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] a2a5fb2d-62e0-4809-a01f-f1df66dca58f/a2a5fb2d-62e0-4809-a01f-f1df66dca58f.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.609231] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51a82afa-5872-40c9-8d01-a02934569b10 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.624125] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "1e7b8237-34ea-479e-b5a7-b1846661d61d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.624125] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "1e7b8237-34ea-479e-b5a7-b1846661d61d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 915.624125] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "1e7b8237-34ea-479e-b5a7-b1846661d61d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.624313] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "1e7b8237-34ea-479e-b5a7-b1846661d61d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 915.624536] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "1e7b8237-34ea-479e-b5a7-b1846661d61d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 915.626892] env[65522]: INFO nova.compute.manager [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Terminating instance [ 915.636773] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 915.636773] env[65522]: value = "task-5114298" [ 915.636773] env[65522]: _type = "Task" [ 915.636773] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.645188] env[65522]: DEBUG nova.compute.manager [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 915.648111] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114298, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.649289] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5074d493-896f-4671-a4b6-50a39465db2b tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.029s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 915.654252] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.233s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 915.656470] env[65522]: INFO nova.compute.claims [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.671497] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114293, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.095212} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.671974] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 9c39b722-56ff-44fa-8f66-3e3432645a68/9c39b722-56ff-44fa-8f66-3e3432645a68.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 915.672150] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.675066] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a8a8269d-d975-4d7b-a079-3c5bfd321062 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.678218] env[65522]: DEBUG oslo_concurrency.lockutils [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Releasing lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 915.678218] env[65522]: DEBUG nova.compute.manager [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Received event network-changed-e12c2307-4963-4a5a-af0f-19369217f430 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 915.678218] env[65522]: DEBUG nova.compute.manager [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Refreshing instance network info cache due to event network-changed-e12c2307-4963-4a5a-af0f-19369217f430. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 915.678414] env[65522]: DEBUG oslo_concurrency.lockutils [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Acquiring lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.678478] env[65522]: DEBUG oslo_concurrency.lockutils [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Acquired lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 915.678656] env[65522]: DEBUG nova.network.neutron [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Refreshing network info cache for port e12c2307-4963-4a5a-af0f-19369217f430 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 915.683025] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 915.683025] env[65522]: value = "task-5114299" [ 915.683025] env[65522]: _type = "Task" [ 915.683025] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.694210] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114299, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.705723] env[65522]: DEBUG oslo_vmware.api [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Task: {'id': task-5114297, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.243257} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.706377] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.706469] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 915.707320] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 915.707320] env[65522]: INFO nova.compute.manager [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Took 1.69 seconds to destroy the instance on the hypervisor. [ 915.707320] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 915.707577] env[65522]: DEBUG nova.compute.manager [-] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 915.707577] env[65522]: DEBUG nova.network.neutron [-] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 915.708037] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 915.709029] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 915.709029] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 916.009441] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 916.049641] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: c93b52c3-1143-44d7-8c10-ac58bc25f541] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 916.063025] env[65522]: DEBUG nova.compute.manager [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 916.063025] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 916.063025] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114294, 'name': Destroy_Task, 'duration_secs': 0.855283} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.063439] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c1198b-7e3e-4fd3-a2d5-c8dffb0e3738 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.067741] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Destroyed the VM [ 916.068697] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 916.069276] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-57f9ae2b-04c0-4285-be76-cd1b056a7937 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.082232] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 916.084084] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb879d5c-9c64-45b0-8f79-6865cdd6c8c3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.086317] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 916.086317] env[65522]: value = "task-5114300" [ 916.086317] env[65522]: _type = "Task" [ 916.086317] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.094301] env[65522]: DEBUG oslo_vmware.api [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 916.094301] env[65522]: value = "task-5114301" [ 916.094301] env[65522]: _type = "Task" [ 916.094301] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.095792] env[65522]: DEBUG nova.network.neutron [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Successfully created port: c2eab4eb-5a45-4c96-a145-bea8b314c82d {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 916.109207] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114300, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.118130] env[65522]: DEBUG oslo_vmware.api [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114301, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.131805] env[65522]: DEBUG nova.compute.manager [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 916.131998] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 916.133406] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84eeb0e1-cb71-4338-8e98-97dc00a0daae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.151699] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 916.157520] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114298, 'name': ReconfigVM_Task, 'duration_secs': 0.427571} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.157830] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d40e9b7a-6150-42af-b6c1-767964824769 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.162528] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Reconfigured VM instance instance-00000044 to attach disk [datastore1] a2a5fb2d-62e0-4809-a01f-f1df66dca58f/a2a5fb2d-62e0-4809-a01f-f1df66dca58f.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.166927] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-91ca1c78-e8ec-48df-858c-f690d1ce1fe1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.174438] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 916.174438] env[65522]: value = "task-5114302" [ 916.174438] env[65522]: _type = "Task" [ 916.174438] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.176653] env[65522]: DEBUG oslo_vmware.api [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 916.176653] env[65522]: value = "task-5114303" [ 916.176653] env[65522]: _type = "Task" [ 916.176653] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.185064] env[65522]: WARNING neutronclient.v2_0.client [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 916.185064] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 916.185064] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 916.211997] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114302, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.218903] env[65522]: DEBUG oslo_vmware.api [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114303, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.224241] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114299, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083513} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.225093] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 916.226497] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824d7857-7ae1-4ea8-8264-621acbbd03fb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.255129] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 9c39b722-56ff-44fa-8f66-3e3432645a68/9c39b722-56ff-44fa-8f66-3e3432645a68.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.255891] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f63a7cde-ea26-4fd4-8e6a-b829df865cc9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.279327] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 916.279327] env[65522]: value = "task-5114304" [ 916.279327] env[65522]: _type = "Task" [ 916.279327] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.289950] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114304, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.581124] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: dfe11c4f-a6e3-4d50-9502-bcbbc5de7bb7] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 916.613105] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114300, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.622292] env[65522]: DEBUG oslo_vmware.api [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114301, 'name': PowerOffVM_Task, 'duration_secs': 0.524426} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.622425] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 916.622679] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 916.623101] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-98e27eed-d62b-46ca-9a6d-5deb8d4c7cca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.664947] env[65522]: DEBUG nova.compute.manager [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 916.689870] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114302, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.104231] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 57d32fb8-885e-44ba-967b-69f4bd1f5744] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 917.106712] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 917.106712] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 917.106712] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Deleting the datastore file [datastore1] 75c8848b-aea9-43f1-8697-9224050d1fef {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 917.106712] env[65522]: DEBUG oslo_vmware.api [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114303, 'name': PowerOffVM_Task, 'duration_secs': 0.268646} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.113407] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-73c50ba1-eaed-465c-b211-781b6144ddee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.114905] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 917.115100] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 917.115663] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c20f7477-b850-4720-85cb-26d1ea6cd422 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.125198] env[65522]: DEBUG oslo_vmware.api [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114300, 'name': RemoveSnapshot_Task, 'duration_secs': 0.819918} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.132119] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 917.132457] env[65522]: INFO nova.compute.manager [None req-9bef0b86-ceb5-4758-bade-3843c34d617f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Took 20.94 seconds to snapshot the instance on the hypervisor. [ 917.137308] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114304, 'name': ReconfigVM_Task, 'duration_secs': 0.659467} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.137308] env[65522]: DEBUG oslo_vmware.api [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for the task: (returnval){ [ 917.137308] env[65522]: value = "task-5114306" [ 917.137308] env[65522]: _type = "Task" [ 917.137308] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.137308] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 9c39b722-56ff-44fa-8f66-3e3432645a68/9c39b722-56ff-44fa-8f66-3e3432645a68.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.140971] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-94bac7de-c8b5-47e7-a0f3-7f4a87eff1c5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.158595] env[65522]: DEBUG nova.virt.hardware [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 917.159239] env[65522]: DEBUG nova.virt.hardware [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 917.159687] env[65522]: DEBUG nova.virt.hardware [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 917.159906] env[65522]: DEBUG nova.virt.hardware [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 917.160287] env[65522]: DEBUG nova.virt.hardware [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 917.160287] env[65522]: DEBUG nova.virt.hardware [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 917.161032] env[65522]: DEBUG nova.virt.hardware [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.161032] env[65522]: DEBUG nova.virt.hardware [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 917.161497] env[65522]: DEBUG nova.virt.hardware [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 917.161745] env[65522]: DEBUG nova.virt.hardware [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 917.163381] env[65522]: DEBUG nova.virt.hardware [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 917.163938] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0feb21a1-afcb-4c55-a8cd-a781800bb198 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.182679] env[65522]: DEBUG oslo_vmware.api [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114306, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.183111] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 917.183111] env[65522]: value = "task-5114308" [ 917.183111] env[65522]: _type = "Task" [ 917.183111] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.199998] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd8fde4-410c-4819-9fe7-6d40b82799d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.207437] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 917.207437] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 917.207437] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleting the datastore file [datastore2] 1e7b8237-34ea-479e-b5a7-b1846661d61d {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 917.208327] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f8f79de-1c4c-4d02-b7e4-970b73eb2f85 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.216075] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114302, 'name': Rename_Task, 'duration_secs': 0.519752} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.216819] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 917.217194] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-107d6a3a-dbf0-481f-995b-c19cfc21161c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.235835] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114308, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.240904] env[65522]: DEBUG oslo_vmware.api [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 917.240904] env[65522]: value = "task-5114309" [ 917.240904] env[65522]: _type = "Task" [ 917.240904] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.250145] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 917.250145] env[65522]: value = "task-5114310" [ 917.250145] env[65522]: _type = "Task" [ 917.250145] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.260762] env[65522]: DEBUG oslo_vmware.api [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114309, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.270965] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114310, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.569636] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f0df81-d39e-4fc2-badb-ace9e3174c69 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.581312] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9c6afc-3670-4923-b8cb-d094e3e877b6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.618755] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 917.618959] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Cleaning up deleted instances with incomplete migration {{(pid=65522) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11971}} [ 917.620701] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03bd906-38f7-450a-910b-7afce517ca1e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.630775] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8873981e-4712-42aa-b3e4-79e3c34382dc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.650499] env[65522]: DEBUG nova.compute.provider_tree [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.663790] env[65522]: DEBUG oslo_vmware.api [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114306, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.701875] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114308, 'name': Rename_Task, 'duration_secs': 0.314091} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.702424] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 917.702503] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-872a5578-6f59-4845-a3d7-cfaaa0e0c398 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.710070] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 917.710070] env[65522]: value = "task-5114311" [ 917.710070] env[65522]: _type = "Task" [ 917.710070] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.720069] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114311, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.752893] env[65522]: DEBUG oslo_vmware.api [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114309, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.45824} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.756543] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.756739] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 917.756968] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 917.757247] env[65522]: INFO nova.compute.manager [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Took 1.63 seconds to destroy the instance on the hypervisor. [ 917.757437] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 917.757704] env[65522]: DEBUG nova.compute.manager [-] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 917.757847] env[65522]: DEBUG nova.network.neutron [-] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 917.758122] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 917.758787] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.759079] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.767152] env[65522]: DEBUG nova.network.neutron [-] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 917.773058] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114310, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.124714] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.158225] env[65522]: DEBUG nova.scheduler.client.report [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 918.169594] env[65522]: DEBUG oslo_vmware.api [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Task: {'id': task-5114306, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.698135} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.170311] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 918.170604] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 918.170866] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 918.171169] env[65522]: INFO nova.compute.manager [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Took 2.11 seconds to destroy the instance on the hypervisor. [ 918.172041] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 918.172041] env[65522]: DEBUG nova.compute.manager [-] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 918.172041] env[65522]: DEBUG nova.network.neutron [-] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 918.172326] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 918.173076] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 918.173466] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 918.221947] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114311, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.263101] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114310, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.269730] env[65522]: INFO nova.compute.manager [-] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Took 2.56 seconds to deallocate network for instance. [ 918.564068] env[65522]: DEBUG nova.network.neutron [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Successfully updated port: c2eab4eb-5a45-4c96-a145-bea8b314c82d {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 918.585212] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 918.585600] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 918.626320] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 918.671747] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.017s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 918.672880] env[65522]: DEBUG nova.compute.manager [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 918.677667] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 5.177s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 918.724730] env[65522]: DEBUG oslo_vmware.api [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114311, 'name': PowerOnVM_Task, 'duration_secs': 0.516521} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.726970] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 918.729132] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 918.730172] env[65522]: INFO nova.compute.manager [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Took 17.02 seconds to spawn the instance on the hypervisor. [ 918.730411] env[65522]: DEBUG nova.compute.manager [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 918.732186] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f22a618-34b8-4a94-adb0-e28699e72fc7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.766206] env[65522]: DEBUG oslo_vmware.api [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114310, 'name': PowerOnVM_Task, 'duration_secs': 1.257368} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.766687] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 918.767138] env[65522]: INFO nova.compute.manager [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Took 13.09 seconds to spawn the instance on the hypervisor. [ 918.767345] env[65522]: DEBUG nova.compute.manager [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 918.768262] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff999b45-9d35-46a8-b65b-b2cbd4de0d98 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.783048] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 919.070939] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.070939] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquired lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 919.070939] env[65522]: DEBUG nova.network.neutron [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 919.183589] env[65522]: DEBUG nova.compute.utils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 919.186582] env[65522]: DEBUG nova.objects.instance [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lazy-loading 'migration_context' on Instance uuid 5e227f6c-9f13-416f-8e6d-2f7d931619fd {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.190476] env[65522]: DEBUG nova.compute.manager [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 919.190912] env[65522]: DEBUG nova.network.neutron [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 919.191055] env[65522]: WARNING neutronclient.v2_0.client [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 919.191527] env[65522]: WARNING neutronclient.v2_0.client [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 919.191864] env[65522]: WARNING openstack [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 919.192266] env[65522]: WARNING openstack [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 919.258292] env[65522]: INFO nova.compute.manager [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Took 45.16 seconds to build instance. [ 919.293920] env[65522]: INFO nova.compute.manager [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Took 26.84 seconds to build instance. [ 919.431464] env[65522]: WARNING neutronclient.v2_0.client [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 919.431464] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 919.434287] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 919.526331] env[65522]: DEBUG nova.policy [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a97b2f232f7844c993dccdc386ce3b2c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9614f18e01594bf083e619c11760fe36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 919.577271] env[65522]: WARNING openstack [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 919.577666] env[65522]: WARNING openstack [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 919.690843] env[65522]: DEBUG nova.compute.manager [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 919.762260] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d5d465b8-041d-45b0-9695-f0bd492dd18a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "9c39b722-56ff-44fa-8f66-3e3432645a68" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.686s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 919.801281] env[65522]: DEBUG oslo_concurrency.lockutils [None req-88b33699-9c25-477f-ad68-e53b68ecee12 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.365s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 919.804818] env[65522]: DEBUG nova.network.neutron [-] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 919.846109] env[65522]: DEBUG nova.network.neutron [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Updated VIF entry in instance network info cache for port e12c2307-4963-4a5a-af0f-19369217f430. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 919.846109] env[65522]: DEBUG nova.network.neutron [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Updating instance_info_cache with network_info: [{"id": "e12c2307-4963-4a5a-af0f-19369217f430", "address": "fa:16:3e:9f:b8:7d", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape12c2307-49", "ovs_interfaceid": "e12c2307-4963-4a5a-af0f-19369217f430", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 919.875750] env[65522]: DEBUG nova.network.neutron [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 919.913015] env[65522]: DEBUG nova.network.neutron [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Successfully created port: 1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 919.947018] env[65522]: DEBUG nova.network.neutron [-] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 919.982376] env[65522]: WARNING openstack [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 919.982838] env[65522]: WARNING openstack [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.130751] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7885f858-6e9f-4ddb-beb5-b9ef7ca56256 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.140629] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f64f229-4de2-45a8-b8fb-f353f70c2cfc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.175137] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f5a3ced-af42-4e7d-a276-dfadde4e7c1a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.185041] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae48940-8f94-4883-bee6-08882e69e5de {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.206443] env[65522]: DEBUG nova.compute.provider_tree [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.307666] env[65522]: INFO nova.compute.manager [-] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Took 2.55 seconds to deallocate network for instance. [ 920.348688] env[65522]: DEBUG oslo_concurrency.lockutils [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Releasing lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 920.348965] env[65522]: DEBUG nova.compute.manager [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Received event network-changed-e12c2307-4963-4a5a-af0f-19369217f430 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 920.349146] env[65522]: DEBUG nova.compute.manager [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Refreshing instance network info cache due to event network-changed-e12c2307-4963-4a5a-af0f-19369217f430. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 920.349360] env[65522]: DEBUG oslo_concurrency.lockutils [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Acquiring lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.349501] env[65522]: DEBUG oslo_concurrency.lockutils [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Acquired lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 920.349677] env[65522]: DEBUG nova.network.neutron [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Refreshing network info cache for port e12c2307-4963-4a5a-af0f-19369217f430 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 920.399011] env[65522]: WARNING neutronclient.v2_0.client [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 920.399680] env[65522]: WARNING openstack [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.400047] env[65522]: WARNING openstack [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.448434] env[65522]: INFO nova.compute.manager [-] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Took 2.28 seconds to deallocate network for instance. [ 920.582962] env[65522]: DEBUG nova.compute.manager [req-dc2df9df-b3ec-49ed-a864-3645bc30233e req-03eed517-a551-43da-a633-df59128289da service nova] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Received event network-vif-deleted-4463f170-2656-4332-b1a3-0b0a491984da {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 920.691947] env[65522]: DEBUG nova.network.neutron [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Updating instance_info_cache with network_info: [{"id": "c2eab4eb-5a45-4c96-a145-bea8b314c82d", "address": "fa:16:3e:22:77:b4", "network": {"id": "f49e9b61-6927-4a5b-a12d-f54f408d42b4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1397158171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a80f979f3dc0477e9462b47f7aa87f14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2eab4eb-5a", "ovs_interfaceid": "c2eab4eb-5a45-4c96-a145-bea8b314c82d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 920.703133] env[65522]: DEBUG nova.compute.manager [req-df590004-1fad-42aa-aef1-eefea01a73cd req-c09f2608-172a-4aa8-ba97-65ef70c3970e service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Received event network-vif-plugged-c2eab4eb-5a45-4c96-a145-bea8b314c82d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 920.703133] env[65522]: DEBUG oslo_concurrency.lockutils [req-df590004-1fad-42aa-aef1-eefea01a73cd req-c09f2608-172a-4aa8-ba97-65ef70c3970e service nova] Acquiring lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 920.703133] env[65522]: DEBUG oslo_concurrency.lockutils [req-df590004-1fad-42aa-aef1-eefea01a73cd req-c09f2608-172a-4aa8-ba97-65ef70c3970e service nova] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 920.703133] env[65522]: DEBUG oslo_concurrency.lockutils [req-df590004-1fad-42aa-aef1-eefea01a73cd req-c09f2608-172a-4aa8-ba97-65ef70c3970e service nova] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 920.703133] env[65522]: DEBUG nova.compute.manager [req-df590004-1fad-42aa-aef1-eefea01a73cd req-c09f2608-172a-4aa8-ba97-65ef70c3970e service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] No waiting events found dispatching network-vif-plugged-c2eab4eb-5a45-4c96-a145-bea8b314c82d {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 920.703133] env[65522]: WARNING nova.compute.manager [req-df590004-1fad-42aa-aef1-eefea01a73cd req-c09f2608-172a-4aa8-ba97-65ef70c3970e service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Received unexpected event network-vif-plugged-c2eab4eb-5a45-4c96-a145-bea8b314c82d for instance with vm_state building and task_state spawning. [ 920.709157] env[65522]: DEBUG nova.compute.manager [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 920.712305] env[65522]: DEBUG nova.scheduler.client.report [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 920.738858] env[65522]: DEBUG nova.virt.hardware [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 920.738858] env[65522]: DEBUG nova.virt.hardware [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 920.738858] env[65522]: DEBUG nova.virt.hardware [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 920.738858] env[65522]: DEBUG nova.virt.hardware [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 920.738858] env[65522]: DEBUG nova.virt.hardware [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 920.738858] env[65522]: DEBUG nova.virt.hardware [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 920.738858] env[65522]: DEBUG nova.virt.hardware [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.738858] env[65522]: DEBUG nova.virt.hardware [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 920.738858] env[65522]: DEBUG nova.virt.hardware [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 920.738858] env[65522]: DEBUG nova.virt.hardware [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 920.739517] env[65522]: DEBUG nova.virt.hardware [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 920.740424] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6177ad5-7df5-4c73-a470-f3441411a2de {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.749811] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342e6c86-4565-4a05-8da8-e9c75b3782ed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.814391] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 920.853117] env[65522]: WARNING neutronclient.v2_0.client [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 920.854123] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.855207] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.955757] env[65522]: DEBUG oslo_concurrency.lockutils [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 921.041648] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 921.042100] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 921.195874] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Releasing lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 921.196359] env[65522]: DEBUG nova.compute.manager [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Instance network_info: |[{"id": "c2eab4eb-5a45-4c96-a145-bea8b314c82d", "address": "fa:16:3e:22:77:b4", "network": {"id": "f49e9b61-6927-4a5b-a12d-f54f408d42b4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1397158171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a80f979f3dc0477e9462b47f7aa87f14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2eab4eb-5a", "ovs_interfaceid": "c2eab4eb-5a45-4c96-a145-bea8b314c82d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 921.196892] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:77:b4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9d39252e-42ef-4252-98d3-62af5a0d109d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2eab4eb-5a45-4c96-a145-bea8b314c82d', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.205444] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Creating folder: Project (a80f979f3dc0477e9462b47f7aa87f14). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 921.206314] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d367c0d-fda8-4a8f-beec-e167a65bbe3a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.224354] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Created folder: Project (a80f979f3dc0477e9462b47f7aa87f14) in parent group-v994660. [ 921.225206] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Creating folder: Instances. Parent ref: group-v994844. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 921.226281] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d0860dfb-c60b-4d81-86d5-1e1f9a2c6f7e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.240956] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Created folder: Instances in parent group-v994844. [ 921.241196] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 921.241385] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 921.241640] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-376becb5-0f43-4fd0-94db-78590009c966 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.269536] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 921.269536] env[65522]: value = "task-5114314" [ 921.269536] env[65522]: _type = "Task" [ 921.269536] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.287963] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114314, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.508190] env[65522]: WARNING neutronclient.v2_0.client [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.509351] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 921.509814] env[65522]: WARNING openstack [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 921.673761] env[65522]: DEBUG nova.network.neutron [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Successfully updated port: 1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 921.728360] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.051s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 921.737776] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.998s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 921.740492] env[65522]: INFO nova.compute.claims [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.790274] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114314, 'name': CreateVM_Task, 'duration_secs': 0.432438} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.790492] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 921.791067] env[65522]: WARNING neutronclient.v2_0.client [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 921.791569] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.791641] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 921.791944] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 921.792546] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14fbbbed-f654-46f7-909d-a7d78c23cd5c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.799829] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 921.799829] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527329a6-f512-eef0-aa52-262d7b8fa3cb" [ 921.799829] env[65522]: _type = "Task" [ 921.799829] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.810619] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527329a6-f512-eef0-aa52-262d7b8fa3cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.832535] env[65522]: DEBUG nova.network.neutron [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Updated VIF entry in instance network info cache for port e12c2307-4963-4a5a-af0f-19369217f430. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 921.832982] env[65522]: DEBUG nova.network.neutron [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Updating instance_info_cache with network_info: [{"id": "e12c2307-4963-4a5a-af0f-19369217f430", "address": "fa:16:3e:9f:b8:7d", "network": {"id": "ccdaa477-c168-4042-a74e-0e1c675a9a4b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2135249252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b8112ba49034e6fabcb3fbbd46edf41", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b033f4d-2e92-4702-add6-410a29d3f251", "external-id": "nsx-vlan-transportzone-649", "segmentation_id": 649, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape12c2307-49", "ovs_interfaceid": "e12c2307-4963-4a5a-af0f-19369217f430", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 921.840060] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Acquiring lock "9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 921.840060] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Lock "9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 922.177109] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "refresh_cache-4aca4e87-b923-49e4-88d8-7ff51bcbe993" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.177363] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "refresh_cache-4aca4e87-b923-49e4-88d8-7ff51bcbe993" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 922.177934] env[65522]: DEBUG nova.network.neutron [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 922.314841] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527329a6-f512-eef0-aa52-262d7b8fa3cb, 'name': SearchDatastore_Task, 'duration_secs': 0.015685} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.315549] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 922.315909] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.316218] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.316414] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 922.316621] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.316931] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-acb7f690-e5aa-416a-a5fb-e1aabc7f5e51 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.331864] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.331864] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 922.331864] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab136ccd-deb0-4b26-8533-16c83b28c729 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.340087] env[65522]: DEBUG oslo_concurrency.lockutils [req-f3b6f26b-c45e-48dd-8280-714c61ae7927 req-7c04b427-379a-4625-baba-3fd9ab385987 service nova] Releasing lock "refresh_cache-1eebb320-30e2-4d8b-b1fd-19e31fb35b77" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 922.341209] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 922.341209] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52baed64-872c-7a8d-b4df-ca88cac478c3" [ 922.341209] env[65522]: _type = "Task" [ 922.341209] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.341683] env[65522]: DEBUG nova.compute.manager [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 922.361233] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52baed64-872c-7a8d-b4df-ca88cac478c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.438320] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "4264998b-949d-4b01-8832-fb1df6214403" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 922.438558] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "4264998b-949d-4b01-8832-fb1df6214403" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 922.438779] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "4264998b-949d-4b01-8832-fb1df6214403-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 922.438956] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "4264998b-949d-4b01-8832-fb1df6214403-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 922.439161] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "4264998b-949d-4b01-8832-fb1df6214403-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 922.442128] env[65522]: INFO nova.compute.manager [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Terminating instance [ 922.485667] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "84d4c689-b437-41a5-bb62-d144a30c6219" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 922.486290] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "84d4c689-b437-41a5-bb62-d144a30c6219" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 922.683688] env[65522]: WARNING openstack [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 922.683856] env[65522]: WARNING openstack [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 922.757199] env[65522]: DEBUG nova.network.neutron [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 922.787102] env[65522]: WARNING openstack [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 922.787102] env[65522]: WARNING openstack [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 922.858157] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52baed64-872c-7a8d-b4df-ca88cac478c3, 'name': SearchDatastore_Task, 'duration_secs': 0.020097} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.868602] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4a6aebf-0f06-4cc5-8855-3542c728bf4f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.876153] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 922.876153] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e04da3-43cc-2a4f-63ca-6be4531589ba" [ 922.876153] env[65522]: _type = "Task" [ 922.876153] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.880235] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 922.885955] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e04da3-43cc-2a4f-63ca-6be4531589ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.913259] env[65522]: WARNING neutronclient.v2_0.client [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 922.914403] env[65522]: WARNING openstack [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 922.915222] env[65522]: WARNING openstack [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 922.945983] env[65522]: DEBUG nova.compute.manager [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 922.946312] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 922.947202] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e96297b-7b0a-4a3c-b83b-17b1a3e99d9b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.956016] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 922.959392] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6ef57939-6c9f-401a-9150-7c2667a0e546 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.994163] env[65522]: DEBUG nova.compute.manager [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 923.056385] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 923.056507] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 923.056676] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleting the datastore file [datastore1] 4264998b-949d-4b01-8832-fb1df6214403 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 923.056935] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c21c701-3f8a-4878-987c-ac6f36db67c0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.067054] env[65522]: DEBUG oslo_vmware.api [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 923.067054] env[65522]: value = "task-5114316" [ 923.067054] env[65522]: _type = "Task" [ 923.067054] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.082258] env[65522]: DEBUG oslo_vmware.api [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.148490] env[65522]: DEBUG nova.network.neutron [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Updating instance_info_cache with network_info: [{"id": "1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2", "address": "fa:16:3e:73:50:60", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cc60bc4-0f", "ovs_interfaceid": "1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 923.291395] env[65522]: INFO nova.compute.manager [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Swapping old allocation on dict_keys(['9a0d1ac8-ed26-4657-8899-fbceab29db0d']) held by migration ed83be76-1d3a-4a25-85ce-9cf749f05abc for instance [ 923.311707] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb587a2-5fec-493a-aab3-19ca99c3e7d0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.324646] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a906318e-5813-44ee-a145-c6414a47b6e0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.331327] env[65522]: DEBUG nova.scheduler.client.report [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Overwriting current allocation {'allocations': {'9a0d1ac8-ed26-4657-8899-fbceab29db0d': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 92}}, 'project_id': '503adb79a1f74f1b98e1f06ec54f1584', 'user_id': '1103ba172d754007bb2b7339e1d674f5', 'consumer_generation': 1} on consumer 5e227f6c-9f13-416f-8e6d-2f7d931619fd {{(pid=65522) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 923.386723] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4064c6fb-d6b6-4ca1-9b06-f8aaf1046152 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.401589] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e04da3-43cc-2a4f-63ca-6be4531589ba, 'name': SearchDatastore_Task, 'duration_secs': 0.043967} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.406797] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 923.406797] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 5ce4a286-efd1-4bbc-a23b-931c6701cfe4/5ce4a286-efd1-4bbc-a23b-931c6701cfe4.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 923.408403] env[65522]: WARNING neutronclient.v2_0.client [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 923.412250] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-852ec075-46ef-4552-94c2-044c9a7a06cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.419146] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c280534-371e-466f-b203-1c79219e0ee8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.429825] env[65522]: DEBUG nova.compute.manager [req-d1f163d2-d829-4833-a90a-8ef1548d2695 req-e83fee79-c248-4dac-a91b-f902c43b9881 service nova] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Received event network-vif-deleted-37951fc0-2513-48ca-8e6a-42be7de65465 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 923.442775] env[65522]: DEBUG nova.compute.provider_tree [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.446273] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 923.446273] env[65522]: value = "task-5114317" [ 923.446273] env[65522]: _type = "Task" [ 923.446273] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.458409] env[65522]: DEBUG nova.compute.manager [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Received event network-changed-c2eab4eb-5a45-4c96-a145-bea8b314c82d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 923.458885] env[65522]: DEBUG nova.compute.manager [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Refreshing instance network info cache due to event network-changed-c2eab4eb-5a45-4c96-a145-bea8b314c82d. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 923.461420] env[65522]: DEBUG oslo_concurrency.lockutils [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Acquiring lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.461650] env[65522]: DEBUG oslo_concurrency.lockutils [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Acquired lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 923.461897] env[65522]: DEBUG nova.network.neutron [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Refreshing network info cache for port c2eab4eb-5a45-4c96-a145-bea8b314c82d {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 923.466975] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114317, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.492242] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.493710] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquired lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 923.493710] env[65522]: DEBUG nova.network.neutron [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 923.524521] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 923.578604] env[65522]: DEBUG oslo_vmware.api [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114316, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.39419} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.578835] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.579104] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 923.579195] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 923.579369] env[65522]: INFO nova.compute.manager [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Took 0.63 seconds to destroy the instance on the hypervisor. [ 923.579694] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 923.579820] env[65522]: DEBUG nova.compute.manager [-] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 923.579916] env[65522]: DEBUG nova.network.neutron [-] [instance: 4264998b-949d-4b01-8832-fb1df6214403] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 923.580206] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 923.581153] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 923.581153] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 923.652325] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 923.656172] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "refresh_cache-4aca4e87-b923-49e4-88d8-7ff51bcbe993" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 923.656622] env[65522]: DEBUG nova.compute.manager [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Instance network_info: |[{"id": "1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2", "address": "fa:16:3e:73:50:60", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cc60bc4-0f", "ovs_interfaceid": "1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 923.657573] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:50:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10b81051-1eb1-406b-888c-4548c470c77e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 923.666694] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Creating folder: Project (9614f18e01594bf083e619c11760fe36). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 923.667184] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-abbe8bd1-0f89-44e2-b920-d2fd557b9fa1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.682353] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Created folder: Project (9614f18e01594bf083e619c11760fe36) in parent group-v994660. [ 923.682558] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Creating folder: Instances. Parent ref: group-v994847. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 923.682832] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-361c34ab-fcdf-45af-87ec-1fe1197b2f96 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.695603] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Created folder: Instances in parent group-v994847. [ 923.696015] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 923.696346] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 923.696598] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aff284dd-360e-4d1b-8742-0e42d0ab3f1b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.720094] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 923.720094] env[65522]: value = "task-5114320" [ 923.720094] env[65522]: _type = "Task" [ 923.720094] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.740831] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114320, 'name': CreateVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.948727] env[65522]: DEBUG nova.scheduler.client.report [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 923.966325] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114317, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.970217] env[65522]: WARNING neutronclient.v2_0.client [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 923.970881] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 923.971667] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 923.995557] env[65522]: WARNING neutronclient.v2_0.client [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 923.996488] env[65522]: WARNING openstack [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 923.996900] env[65522]: WARNING openstack [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 924.165305] env[65522]: WARNING openstack [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 924.165903] env[65522]: WARNING openstack [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 924.185119] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 924.185672] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 924.234888] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Acquiring lock "42a08cfb-d865-4967-a086-370a2ca98b7d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 924.234888] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Lock "42a08cfb-d865-4967-a086-370a2ca98b7d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 924.234888] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Acquiring lock "42a08cfb-d865-4967-a086-370a2ca98b7d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 924.234888] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Lock "42a08cfb-d865-4967-a086-370a2ca98b7d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 924.234888] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Lock "42a08cfb-d865-4967-a086-370a2ca98b7d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 924.237357] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114320, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.238987] env[65522]: INFO nova.compute.manager [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Terminating instance [ 924.339834] env[65522]: WARNING neutronclient.v2_0.client [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 924.340667] env[65522]: WARNING openstack [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 924.341059] env[65522]: WARNING openstack [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 924.377675] env[65522]: WARNING neutronclient.v2_0.client [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 924.378237] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 924.378616] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 924.411227] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "8da2a500-6f0e-4eda-9a92-79510753b3b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 924.411460] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "8da2a500-6f0e-4eda-9a92-79510753b3b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 924.429376] env[65522]: DEBUG nova.network.neutron [-] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 924.453711] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.716s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 924.454238] env[65522]: DEBUG nova.compute.manager [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 924.464118] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.904s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 924.464118] env[65522]: INFO nova.compute.claims [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 924.475839] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114317, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.721787} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.476393] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 5ce4a286-efd1-4bbc-a23b-931c6701cfe4/5ce4a286-efd1-4bbc-a23b-931c6701cfe4.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 924.476639] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 924.476903] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-47d948a0-b483-4dca-a3ee-f2dd365bd2f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.490484] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 924.490484] env[65522]: value = "task-5114321" [ 924.490484] env[65522]: _type = "Task" [ 924.490484] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.505825] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114321, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.591137] env[65522]: DEBUG nova.network.neutron [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance_info_cache with network_info: [{"id": "3f79042c-23dd-4fc0-b3cf-1e2884edabc1", "address": "fa:16:3e:f7:f5:9e", "network": {"id": "491c1959-e42b-49e3-8562-1caf859054c3", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.131", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "596b2b0744b64deb86a3dbe6da5c8894", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc0e97b-b21d-4557-a4d4-fd7e8f973368", "external-id": "nsx-vlan-transportzone-380", "segmentation_id": 380, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f79042c-23", "ovs_interfaceid": "3f79042c-23dd-4fc0-b3cf-1e2884edabc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 924.622486] env[65522]: DEBUG nova.network.neutron [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Updated VIF entry in instance network info cache for port c2eab4eb-5a45-4c96-a145-bea8b314c82d. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 924.622486] env[65522]: DEBUG nova.network.neutron [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Updating instance_info_cache with network_info: [{"id": "c2eab4eb-5a45-4c96-a145-bea8b314c82d", "address": "fa:16:3e:22:77:b4", "network": {"id": "f49e9b61-6927-4a5b-a12d-f54f408d42b4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1397158171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a80f979f3dc0477e9462b47f7aa87f14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2eab4eb-5a", "ovs_interfaceid": "c2eab4eb-5a45-4c96-a145-bea8b314c82d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 924.740532] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114320, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.742808] env[65522]: DEBUG nova.compute.manager [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 924.743068] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 924.743562] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-51a55959-1459-44a7-b8a2-306222124518 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.752821] env[65522]: DEBUG oslo_vmware.api [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Waiting for the task: (returnval){ [ 924.752821] env[65522]: value = "task-5114322" [ 924.752821] env[65522]: _type = "Task" [ 924.752821] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.761755] env[65522]: DEBUG oslo_vmware.api [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5114322, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.916656] env[65522]: DEBUG nova.compute.manager [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 924.936647] env[65522]: INFO nova.compute.manager [-] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Took 1.36 seconds to deallocate network for instance. [ 924.969556] env[65522]: DEBUG nova.compute.utils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 924.973367] env[65522]: DEBUG nova.compute.manager [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 924.973579] env[65522]: DEBUG nova.network.neutron [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 924.973906] env[65522]: WARNING neutronclient.v2_0.client [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 924.974302] env[65522]: WARNING neutronclient.v2_0.client [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 924.974993] env[65522]: WARNING openstack [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 924.975428] env[65522]: WARNING openstack [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 925.000498] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114321, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079511} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.000827] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 925.002497] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc7d36d-b61f-4a18-bfc3-af697ad48229 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.026173] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 5ce4a286-efd1-4bbc-a23b-931c6701cfe4/5ce4a286-efd1-4bbc-a23b-931c6701cfe4.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.029679] env[65522]: DEBUG nova.policy [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd99ebf222a2c4fd295c182b1a14b68bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ea1965645e741f6912422c21417f1ec', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 925.031777] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9a8fc53-bfc6-4f6c-a113-626c6411ae59 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.055375] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 925.055375] env[65522]: value = "task-5114323" [ 925.055375] env[65522]: _type = "Task" [ 925.055375] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.066309] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114323, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.093178] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Releasing lock "refresh_cache-5e227f6c-9f13-416f-8e6d-2f7d931619fd" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 925.093178] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.093371] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-03484458-c365-490a-a18a-1576afc7da17 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.104923] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 925.104923] env[65522]: value = "task-5114324" [ 925.104923] env[65522]: _type = "Task" [ 925.104923] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.115901] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114324, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.124827] env[65522]: DEBUG oslo_concurrency.lockutils [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Releasing lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 925.125246] env[65522]: DEBUG nova.compute.manager [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Received event network-vif-deleted-0750570e-fa7b-4296-a6f5-dec88fcbbab6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 925.126077] env[65522]: DEBUG nova.compute.manager [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Received event network-vif-plugged-1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 925.126077] env[65522]: DEBUG oslo_concurrency.lockutils [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Acquiring lock "4aca4e87-b923-49e4-88d8-7ff51bcbe993-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 925.126077] env[65522]: DEBUG oslo_concurrency.lockutils [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Lock "4aca4e87-b923-49e4-88d8-7ff51bcbe993-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 925.126253] env[65522]: DEBUG oslo_concurrency.lockutils [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Lock "4aca4e87-b923-49e4-88d8-7ff51bcbe993-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 925.126760] env[65522]: DEBUG nova.compute.manager [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] No waiting events found dispatching network-vif-plugged-1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 925.127010] env[65522]: WARNING nova.compute.manager [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Received unexpected event network-vif-plugged-1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2 for instance with vm_state building and task_state spawning. [ 925.127174] env[65522]: DEBUG nova.compute.manager [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Received event network-changed-1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 925.127379] env[65522]: DEBUG nova.compute.manager [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Refreshing instance network info cache due to event network-changed-1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 925.127614] env[65522]: DEBUG oslo_concurrency.lockutils [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Acquiring lock "refresh_cache-4aca4e87-b923-49e4-88d8-7ff51bcbe993" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.127805] env[65522]: DEBUG oslo_concurrency.lockutils [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Acquired lock "refresh_cache-4aca4e87-b923-49e4-88d8-7ff51bcbe993" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 925.128253] env[65522]: DEBUG nova.network.neutron [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Refreshing network info cache for port 1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 925.235805] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114320, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.270308] env[65522]: DEBUG oslo_vmware.api [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5114322, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.327633] env[65522]: DEBUG nova.compute.manager [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Stashing vm_state: active {{(pid=65522) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 925.444529] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 925.454078] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 925.475259] env[65522]: DEBUG nova.compute.manager [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 925.495655] env[65522]: DEBUG nova.network.neutron [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Successfully created port: 41b80f01-c268-48b3-8886-1273938ac6bd {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 925.569409] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114323, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.615848] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114324, 'name': PowerOffVM_Task, 'duration_secs': 0.361739} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.615848] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 925.616821] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:01:50Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1d8f2c10-9623-4a70-b964-018e648fd761',id=39,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-811894183',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 925.616821] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 925.617062] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 925.617144] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 925.617655] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 925.617655] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 925.617748] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 925.617865] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 925.618038] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 925.618834] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 925.618834] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 925.630015] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97c186f0-a5a7-491e-beb6-ed7d53b84ae2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.642912] env[65522]: WARNING neutronclient.v2_0.client [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 925.642912] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 925.644022] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 925.657473] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 925.657473] env[65522]: value = "task-5114325" [ 925.657473] env[65522]: _type = "Task" [ 925.657473] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.673373] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114325, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.738080] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114320, 'name': CreateVM_Task, 'duration_secs': 1.756736} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.738080] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 925.738080] env[65522]: WARNING neutronclient.v2_0.client [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 925.738928] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.738928] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 925.738928] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 925.739287] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd19950a-0b0c-4a85-8991-94aa53e5ac30 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.745196] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 925.745196] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52938b84-d772-00dd-acc6-b19339a195d5" [ 925.745196] env[65522]: _type = "Task" [ 925.745196] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.761952] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52938b84-d772-00dd-acc6-b19339a195d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.773661] env[65522]: DEBUG oslo_vmware.api [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5114322, 'name': PowerOffVM_Task, 'duration_secs': 0.877613} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.773968] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 925.774198] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 925.774447] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994729', 'volume_id': '78aedba8-1717-4210-afe1-44eb7ae91932', 'name': 'volume-78aedba8-1717-4210-afe1-44eb7ae91932', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '42a08cfb-d865-4967-a086-370a2ca98b7d', 'attached_at': '', 'detached_at': '', 'volume_id': '78aedba8-1717-4210-afe1-44eb7ae91932', 'serial': '78aedba8-1717-4210-afe1-44eb7ae91932'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 925.775462] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b05ba01-8808-4baa-9169-57c48eb506aa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.803858] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c73517c-6bcb-4af3-886c-e69badc4961d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.815278] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb624c71-0149-46d7-802a-c2b68c440528 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.842214] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 925.842626] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 925.849418] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c3019b2-9a7b-4ea3-8142-13417175892a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.856024] env[65522]: DEBUG nova.compute.manager [req-0a7f6892-c164-4835-9cfe-b296ada7b713 req-10c91dec-f110-4621-a7f6-fc24d827ee74 service nova] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Received event network-vif-deleted-a5bdb6f4-7094-4bee-9867-c1d2e683d2f3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 925.857508] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 925.869042] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] The volume has not been displaced from its original location: [datastore2] volume-78aedba8-1717-4210-afe1-44eb7ae91932/volume-78aedba8-1717-4210-afe1-44eb7ae91932.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 925.876662] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Reconfiguring VM instance instance-0000001f to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 925.880056] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-faaf551e-ac70-4503-977b-3afacbce43d5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.904561] env[65522]: DEBUG oslo_vmware.api [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Waiting for the task: (returnval){ [ 925.904561] env[65522]: value = "task-5114326" [ 925.904561] env[65522]: _type = "Task" [ 925.904561] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.919366] env[65522]: DEBUG oslo_vmware.api [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5114326, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.962413] env[65522]: WARNING neutronclient.v2_0.client [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 925.963050] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 925.963412] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 926.064510] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e342348-970f-41f8-82b4-663731f81cea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.074085] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114323, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.077683] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f0619e-0404-453b-a1db-74d01d32e492 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.117570] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584d9f1f-398c-45e1-9ebe-0e252e8e4fbd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.126305] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1664604-507c-4796-9a91-3e26440edd24 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.144101] env[65522]: DEBUG nova.compute.provider_tree [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.174046] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.214664] env[65522]: DEBUG nova.network.neutron [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Updated VIF entry in instance network info cache for port 1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 926.215095] env[65522]: DEBUG nova.network.neutron [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Updating instance_info_cache with network_info: [{"id": "1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2", "address": "fa:16:3e:73:50:60", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cc60bc4-0f", "ovs_interfaceid": "1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 926.257260] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52938b84-d772-00dd-acc6-b19339a195d5, 'name': SearchDatastore_Task, 'duration_secs': 0.193041} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.257638] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 926.257887] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 926.258151] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.258341] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 926.258576] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 926.258900] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab5bc17b-7747-4da8-956a-f3eb13a543a9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.271846] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.272077] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 926.272975] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6645fabf-9dc3-4fa0-a8e8-11da8fd4aa32 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.281701] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 926.281701] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d77218-1ed5-079d-4fbe-a463f58a56c4" [ 926.281701] env[65522]: _type = "Task" [ 926.281701] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.290709] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d77218-1ed5-079d-4fbe-a463f58a56c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.417219] env[65522]: DEBUG oslo_vmware.api [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5114326, 'name': ReconfigVM_Task, 'duration_secs': 0.310111} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.417475] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Reconfigured VM instance instance-0000001f to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 926.422714] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-047d43c7-345c-4b8a-b0c7-3f449fc9b535 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.441183] env[65522]: DEBUG oslo_vmware.api [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Waiting for the task: (returnval){ [ 926.441183] env[65522]: value = "task-5114327" [ 926.441183] env[65522]: _type = "Task" [ 926.441183] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.684168] env[65522]: DEBUG oslo_vmware.api [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5114327, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.684168] env[65522]: DEBUG nova.compute.manager [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 926.684168] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114323, 'name': ReconfigVM_Task, 'duration_secs': 1.173976} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.684168] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 5ce4a286-efd1-4bbc-a23b-931c6701cfe4/5ce4a286-efd1-4bbc-a23b-931c6701cfe4.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.684168] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09122792-bf42-4aa2-bac3-e081f73b411a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.684168] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 926.684168] env[65522]: value = "task-5114328" [ 926.684168] env[65522]: _type = "Task" [ 926.684168] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.684168] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114328, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.684168] env[65522]: DEBUG nova.scheduler.client.report [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 926.684168] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114325, 'name': ReconfigVM_Task, 'duration_secs': 0.579125} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.684168] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9374e5fc-a843-4b36-9e51-b24b2a40a48e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.694881] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T19:01:50Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1d8f2c10-9623-4a70-b964-018e648fd761',id=39,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-811894183',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 926.695178] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 926.695396] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 926.695612] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 926.696269] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 926.696269] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 926.696269] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.696419] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 926.696585] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 926.696858] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 926.697010] env[65522]: DEBUG nova.virt.hardware [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 926.698317] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc18281e-653e-4826-b1c8-20a1d2009abb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.708103] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 926.708103] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a35cd6-4bc7-f725-71de-7c1608fdf3d9" [ 926.708103] env[65522]: _type = "Task" [ 926.708103] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.715785] env[65522]: DEBUG nova.virt.hardware [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 926.716198] env[65522]: DEBUG nova.virt.hardware [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 926.716322] env[65522]: DEBUG nova.virt.hardware [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 926.716552] env[65522]: DEBUG nova.virt.hardware [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 926.717108] env[65522]: DEBUG nova.virt.hardware [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 926.717108] env[65522]: DEBUG nova.virt.hardware [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 926.717108] env[65522]: DEBUG nova.virt.hardware [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.717326] env[65522]: DEBUG nova.virt.hardware [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 926.717543] env[65522]: DEBUG nova.virt.hardware [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 926.717740] env[65522]: DEBUG nova.virt.hardware [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 926.717965] env[65522]: DEBUG nova.virt.hardware [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 926.719073] env[65522]: DEBUG oslo_concurrency.lockutils [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Releasing lock "refresh_cache-4aca4e87-b923-49e4-88d8-7ff51bcbe993" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 926.719355] env[65522]: DEBUG nova.compute.manager [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Received event network-changed-54206d03-4403-4bca-8ba8-1f4e88682cb6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 926.719575] env[65522]: DEBUG nova.compute.manager [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Refreshing instance network info cache due to event network-changed-54206d03-4403-4bca-8ba8-1f4e88682cb6. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 926.719848] env[65522]: DEBUG oslo_concurrency.lockutils [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Acquiring lock "refresh_cache-9c39b722-56ff-44fa-8f66-3e3432645a68" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.720034] env[65522]: DEBUG oslo_concurrency.lockutils [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Acquired lock "refresh_cache-9c39b722-56ff-44fa-8f66-3e3432645a68" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 926.720278] env[65522]: DEBUG nova.network.neutron [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Refreshing network info cache for port 54206d03-4403-4bca-8ba8-1f4e88682cb6 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 926.726067] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9228c94b-b8bd-42c4-82c7-a7ca3acb368c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.748800] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 926.749089] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 926.750771] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a35cd6-4bc7-f725-71de-7c1608fdf3d9, 'name': SearchDatastore_Task, 'duration_secs': 0.030254} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.759330] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Reconfiguring VM instance instance-00000039 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 926.762618] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-664c7ec7-39b1-42bd-92e0-11badfd20a54 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.779391] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994a165b-6ec2-4613-978d-90035d1a3d32 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.793250] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 926.793250] env[65522]: value = "task-5114329" [ 926.793250] env[65522]: _type = "Task" [ 926.793250] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.809148] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d77218-1ed5-079d-4fbe-a463f58a56c4, 'name': SearchDatastore_Task, 'duration_secs': 0.03283} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.810593] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f01c2934-a695-49a3-b54d-a09323af6a90 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.817667] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114329, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.823294] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 926.823294] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52977686-b61f-859c-8ca8-21fa58e6e757" [ 926.823294] env[65522]: _type = "Task" [ 926.823294] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.834287] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52977686-b61f-859c-8ca8-21fa58e6e757, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.959834] env[65522]: DEBUG oslo_vmware.api [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5114327, 'name': ReconfigVM_Task, 'duration_secs': 0.200483} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.960435] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994729', 'volume_id': '78aedba8-1717-4210-afe1-44eb7ae91932', 'name': 'volume-78aedba8-1717-4210-afe1-44eb7ae91932', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '42a08cfb-d865-4967-a086-370a2ca98b7d', 'attached_at': '', 'detached_at': '', 'volume_id': '78aedba8-1717-4210-afe1-44eb7ae91932', 'serial': '78aedba8-1717-4210-afe1-44eb7ae91932'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 926.960892] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 926.962019] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b1b442-a5f1-4975-9bbe-5d58ff2aec0f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.973936] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 926.977015] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b4f2135a-9237-4575-aa89-75d6ac1f2c2d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.056898] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.057153] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.057340] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Deleting the datastore file [datastore2] 42a08cfb-d865-4967-a086-370a2ca98b7d {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.057643] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14bbd31a-0d3b-4436-8f78-072519f72eeb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.070404] env[65522]: DEBUG oslo_vmware.api [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Waiting for the task: (returnval){ [ 927.070404] env[65522]: value = "task-5114331" [ 927.070404] env[65522]: _type = "Task" [ 927.070404] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.079758] env[65522]: DEBUG oslo_vmware.api [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5114331, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.096104] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114328, 'name': Rename_Task, 'duration_secs': 0.287721} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.096244] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 927.096532] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6049175-9903-4f2f-9237-0eebf6ec9b96 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.105065] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 927.105065] env[65522]: value = "task-5114332" [ 927.105065] env[65522]: _type = "Task" [ 927.105065] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.114275] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114332, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.175068] env[65522]: DEBUG nova.network.neutron [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Successfully updated port: 41b80f01-c268-48b3-8886-1273938ac6bd {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 927.199988] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.737s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 927.200655] env[65522]: DEBUG nova.compute.manager [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 927.205181] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.421s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 927.205181] env[65522]: DEBUG nova.objects.instance [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Lazy-loading 'resources' on Instance uuid 55ab4771-9908-4640-a142-3cb40c0c6ee5 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.226288] env[65522]: DEBUG nova.compute.manager [req-d0c7e980-e479-4708-a1ae-feaaed214207 req-42a55cc7-527b-4f0f-9fc6-27c3b7365a17 service nova] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Received event network-vif-plugged-41b80f01-c268-48b3-8886-1273938ac6bd {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 927.226288] env[65522]: DEBUG oslo_concurrency.lockutils [req-d0c7e980-e479-4708-a1ae-feaaed214207 req-42a55cc7-527b-4f0f-9fc6-27c3b7365a17 service nova] Acquiring lock "2a12a0a3-913e-4ade-85ef-cca7ba508ac9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 927.226288] env[65522]: DEBUG oslo_concurrency.lockutils [req-d0c7e980-e479-4708-a1ae-feaaed214207 req-42a55cc7-527b-4f0f-9fc6-27c3b7365a17 service nova] Lock "2a12a0a3-913e-4ade-85ef-cca7ba508ac9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 927.226620] env[65522]: DEBUG oslo_concurrency.lockutils [req-d0c7e980-e479-4708-a1ae-feaaed214207 req-42a55cc7-527b-4f0f-9fc6-27c3b7365a17 service nova] Lock "2a12a0a3-913e-4ade-85ef-cca7ba508ac9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 927.226620] env[65522]: DEBUG nova.compute.manager [req-d0c7e980-e479-4708-a1ae-feaaed214207 req-42a55cc7-527b-4f0f-9fc6-27c3b7365a17 service nova] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] No waiting events found dispatching network-vif-plugged-41b80f01-c268-48b3-8886-1273938ac6bd {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 927.226695] env[65522]: WARNING nova.compute.manager [req-d0c7e980-e479-4708-a1ae-feaaed214207 req-42a55cc7-527b-4f0f-9fc6-27c3b7365a17 service nova] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Received unexpected event network-vif-plugged-41b80f01-c268-48b3-8886-1273938ac6bd for instance with vm_state building and task_state spawning. [ 927.231357] env[65522]: WARNING neutronclient.v2_0.client [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 927.232091] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.232486] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.252495] env[65522]: DEBUG nova.compute.manager [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 927.306250] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114329, 'name': ReconfigVM_Task, 'duration_secs': 0.219167} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.306563] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Reconfigured VM instance instance-00000039 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 927.307408] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efccedfa-7e9a-43a9-9276-ff83a11fb272 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.337425] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 5e227f6c-9f13-416f-8e6d-2f7d931619fd/5e227f6c-9f13-416f-8e6d-2f7d931619fd.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.341748] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9473b218-e52c-4dee-96d3-61249e347239 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.362876] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52977686-b61f-859c-8ca8-21fa58e6e757, 'name': SearchDatastore_Task, 'duration_secs': 0.014765} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.368806] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 927.369108] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 4aca4e87-b923-49e4-88d8-7ff51bcbe993/4aca4e87-b923-49e4-88d8-7ff51bcbe993.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 927.369869] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 927.369869] env[65522]: value = "task-5114333" [ 927.369869] env[65522]: _type = "Task" [ 927.369869] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.370538] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-68b25f13-2ae7-4338-bbcd-c6fad169ddcb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.390209] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 927.390209] env[65522]: value = "task-5114334" [ 927.390209] env[65522]: _type = "Task" [ 927.390209] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.390209] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114333, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.421343] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.422050] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.497684] env[65522]: WARNING neutronclient.v2_0.client [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 927.498459] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.498892] env[65522]: WARNING openstack [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.580436] env[65522]: DEBUG oslo_vmware.api [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Task: {'id': task-5114331, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151414} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.580718] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.580962] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 927.581184] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 927.581423] env[65522]: INFO nova.compute.manager [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Took 2.84 seconds to destroy the instance on the hypervisor. [ 927.582129] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 927.582129] env[65522]: DEBUG nova.compute.manager [-] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 927.582129] env[65522]: DEBUG nova.network.neutron [-] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 927.582291] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 927.582957] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.583241] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.595086] env[65522]: DEBUG nova.network.neutron [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Updated VIF entry in instance network info cache for port 54206d03-4403-4bca-8ba8-1f4e88682cb6. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 927.595672] env[65522]: DEBUG nova.network.neutron [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Updating instance_info_cache with network_info: [{"id": "54206d03-4403-4bca-8ba8-1f4e88682cb6", "address": "fa:16:3e:72:7b:af", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.240", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54206d03-44", "ovs_interfaceid": "54206d03-4403-4bca-8ba8-1f4e88682cb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 927.618241] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114332, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.623843] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 927.677965] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Acquiring lock "refresh_cache-2a12a0a3-913e-4ade-85ef-cca7ba508ac9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.678266] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Acquired lock "refresh_cache-2a12a0a3-913e-4ade-85ef-cca7ba508ac9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 927.678487] env[65522]: DEBUG nova.network.neutron [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 927.713161] env[65522]: DEBUG nova.compute.utils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 927.717767] env[65522]: DEBUG nova.compute.manager [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 927.717767] env[65522]: DEBUG nova.network.neutron [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 927.717767] env[65522]: WARNING neutronclient.v2_0.client [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 927.717767] env[65522]: WARNING neutronclient.v2_0.client [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 927.719191] env[65522]: WARNING openstack [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.719636] env[65522]: WARNING openstack [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.777579] env[65522]: DEBUG nova.policy [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'be6b0c0be0b14eb49212319102254972', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd3c9f139a53f47d8bfcb22fa53d7b7b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 927.780055] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 927.888806] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114333, 'name': ReconfigVM_Task, 'duration_secs': 0.481413} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.889126] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 5e227f6c-9f13-416f-8e6d-2f7d931619fd/5e227f6c-9f13-416f-8e6d-2f7d931619fd.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.890202] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34143348-7ebc-4b2d-904e-b9bf99cd035f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.909760] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114334, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.927979] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-471d80d8-e202-40b2-b5f2-eccd207181ce {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.954031] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2709fdce-1072-48a3-b023-e827c9c65960 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.979866] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b635254f-4059-498e-81af-dbfa056e227c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.989891] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 927.989891] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-abda8ef3-93d6-48f1-aa31-7247a8ea9842 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.997270] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 927.997270] env[65522]: value = "task-5114335" [ 927.997270] env[65522]: _type = "Task" [ 927.997270] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.009638] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114335, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.098019] env[65522]: DEBUG oslo_concurrency.lockutils [req-88812e53-c2f1-40c9-a0e4-91af21fd0091 req-0dcdc153-b601-4fa3-b57c-e277154603ad service nova] Releasing lock "refresh_cache-9c39b722-56ff-44fa-8f66-3e3432645a68" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 928.121087] env[65522]: DEBUG oslo_vmware.api [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114332, 'name': PowerOnVM_Task, 'duration_secs': 0.631919} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.121988] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 928.122319] env[65522]: INFO nova.compute.manager [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Took 11.46 seconds to spawn the instance on the hypervisor. [ 928.123686] env[65522]: DEBUG nova.compute.manager [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 928.123686] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb611ac7-092b-41ca-b353-3d2fedc2f128 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.156370] env[65522]: DEBUG nova.network.neutron [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Successfully created port: 7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 928.184508] env[65522]: WARNING openstack [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.184508] env[65522]: WARNING openstack [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.232422] env[65522]: DEBUG nova.compute.manager [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 928.277148] env[65522]: DEBUG nova.network.neutron [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 928.323016] env[65522]: WARNING openstack [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.323561] env[65522]: WARNING openstack [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.332797] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb55be84-781f-4cdf-9797-b0d29858a134 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.342432] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e797638f-3dba-48d0-b87a-88601bd5162c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.383665] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c81a1742-6bd7-4dce-9ff7-91050dccb2b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.392839] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5548ceaa-fc3b-41cc-9fac-13383852623c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.411192] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114334, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.993785} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.418698] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 4aca4e87-b923-49e4-88d8-7ff51bcbe993/4aca4e87-b923-49e4-88d8-7ff51bcbe993.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 928.418949] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 928.419472] env[65522]: DEBUG nova.compute.provider_tree [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.421645] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fbb2d5bb-6f86-4c55-97a4-8728fd57e970 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.430634] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 928.430634] env[65522]: value = "task-5114336" [ 928.430634] env[65522]: _type = "Task" [ 928.430634] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.443234] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114336, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.459723] env[65522]: WARNING neutronclient.v2_0.client [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 928.460149] env[65522]: WARNING openstack [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.460779] env[65522]: WARNING openstack [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.514636] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114335, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.571962] env[65522]: DEBUG nova.network.neutron [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Updating instance_info_cache with network_info: [{"id": "41b80f01-c268-48b3-8886-1273938ac6bd", "address": "fa:16:3e:7d:ff:88", "network": {"id": "08114cc1-8575-4999-ac80-0c6d101dc304", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-264300140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ea1965645e741f6912422c21417f1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41b80f01-c2", "ovs_interfaceid": "41b80f01-c268-48b3-8886-1273938ac6bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 928.614925] env[65522]: DEBUG nova.network.neutron [-] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 928.649168] env[65522]: INFO nova.compute.manager [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Took 27.62 seconds to build instance. [ 928.925950] env[65522]: DEBUG nova.scheduler.client.report [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 928.941374] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114336, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.449867} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.942355] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 928.943184] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068a073c-f501-4050-acb9-ec33a556887f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.974682] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 4aca4e87-b923-49e4-88d8-7ff51bcbe993/4aca4e87-b923-49e4-88d8-7ff51bcbe993.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.975494] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6057f633-5624-433b-bea6-ec0c24c14374 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.996941] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 928.996941] env[65522]: value = "task-5114337" [ 928.996941] env[65522]: _type = "Task" [ 928.996941] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.009256] env[65522]: DEBUG oslo_vmware.api [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114335, 'name': PowerOnVM_Task, 'duration_secs': 0.618702} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.013087] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.017894] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114337, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.077688] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Releasing lock "refresh_cache-2a12a0a3-913e-4ade-85ef-cca7ba508ac9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 929.078239] env[65522]: DEBUG nova.compute.manager [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Instance network_info: |[{"id": "41b80f01-c268-48b3-8886-1273938ac6bd", "address": "fa:16:3e:7d:ff:88", "network": {"id": "08114cc1-8575-4999-ac80-0c6d101dc304", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-264300140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ea1965645e741f6912422c21417f1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41b80f01-c2", "ovs_interfaceid": "41b80f01-c268-48b3-8886-1273938ac6bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 929.078914] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:ff:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9875d38f-76e2-416c-bfb7-f18a22b0d8ee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '41b80f01-c268-48b3-8886-1273938ac6bd', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 929.088681] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Creating folder: Project (5ea1965645e741f6912422c21417f1ec). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 929.089695] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-375f2f4c-9df7-4949-920e-fdb63f98cadf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.105381] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Created folder: Project (5ea1965645e741f6912422c21417f1ec) in parent group-v994660. [ 929.105615] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Creating folder: Instances. Parent ref: group-v994850. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 929.105901] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ed362325-2178-4c74-9d49-e024f871fb2e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.119026] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Created folder: Instances in parent group-v994850. [ 929.119470] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 929.120059] env[65522]: INFO nova.compute.manager [-] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Took 1.54 seconds to deallocate network for instance. [ 929.120469] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 929.123119] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0d276190-f7be-4e1d-9488-f3694a8322ed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.145568] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 929.145568] env[65522]: value = "task-5114340" [ 929.145568] env[65522]: _type = "Task" [ 929.145568] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.154745] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f23b1feb-6f8f-449f-8deb-9b5925b2eecb tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.150s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.155059] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114340, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.251546] env[65522]: DEBUG nova.compute.manager [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 929.267132] env[65522]: DEBUG nova.compute.manager [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Received event network-changed-41b80f01-c268-48b3-8886-1273938ac6bd {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 929.267543] env[65522]: DEBUG nova.compute.manager [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Refreshing instance network info cache due to event network-changed-41b80f01-c268-48b3-8886-1273938ac6bd. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 929.267829] env[65522]: DEBUG oslo_concurrency.lockutils [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] Acquiring lock "refresh_cache-2a12a0a3-913e-4ade-85ef-cca7ba508ac9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.267988] env[65522]: DEBUG oslo_concurrency.lockutils [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] Acquired lock "refresh_cache-2a12a0a3-913e-4ade-85ef-cca7ba508ac9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 929.268201] env[65522]: DEBUG nova.network.neutron [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Refreshing network info cache for port 41b80f01-c268-48b3-8886-1273938ac6bd {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 929.281839] env[65522]: DEBUG nova.virt.hardware [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 929.282159] env[65522]: DEBUG nova.virt.hardware [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 929.282450] env[65522]: DEBUG nova.virt.hardware [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 929.282535] env[65522]: DEBUG nova.virt.hardware [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 929.282756] env[65522]: DEBUG nova.virt.hardware [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 929.282802] env[65522]: DEBUG nova.virt.hardware [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 929.282982] env[65522]: DEBUG nova.virt.hardware [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 929.283151] env[65522]: DEBUG nova.virt.hardware [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 929.283333] env[65522]: DEBUG nova.virt.hardware [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 929.283503] env[65522]: DEBUG nova.virt.hardware [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 929.283676] env[65522]: DEBUG nova.virt.hardware [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 929.285399] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ac42bf-e88c-46ff-a580-da303f9c52a1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.296823] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8993d7b3-7d4a-44bc-9671-864f70b21e6a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.432416] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.228s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.434620] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.620s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 929.435270] env[65522]: DEBUG nova.objects.instance [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lazy-loading 'resources' on Instance uuid 1e7b8237-34ea-479e-b5a7-b1846661d61d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 929.476779] env[65522]: INFO nova.scheduler.client.report [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Deleted allocations for instance 55ab4771-9908-4640-a142-3cb40c0c6ee5 [ 929.508833] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114337, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.659141] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114340, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.691920] env[65522]: INFO nova.compute.manager [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Took 0.57 seconds to detach 1 volumes for instance. [ 929.694360] env[65522]: DEBUG nova.compute.manager [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Deleting volume: 78aedba8-1717-4210-afe1-44eb7ae91932 {{(pid=65522) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3319}} [ 929.775411] env[65522]: WARNING neutronclient.v2_0.client [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 929.776205] env[65522]: WARNING openstack [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 929.776554] env[65522]: WARNING openstack [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 929.803769] env[65522]: DEBUG nova.network.neutron [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Successfully updated port: 7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 929.975171] env[65522]: DEBUG nova.compute.manager [req-0c35b74e-acd1-47ec-8c50-a1a3f518c59c req-cebc6420-b8ca-4407-81e3-5f51e527071a service nova] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Received event network-vif-plugged-7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 929.975426] env[65522]: DEBUG oslo_concurrency.lockutils [req-0c35b74e-acd1-47ec-8c50-a1a3f518c59c req-cebc6420-b8ca-4407-81e3-5f51e527071a service nova] Acquiring lock "21715aa5-24d4-423b-92a0-be7cc1f60877-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 929.975625] env[65522]: DEBUG oslo_concurrency.lockutils [req-0c35b74e-acd1-47ec-8c50-a1a3f518c59c req-cebc6420-b8ca-4407-81e3-5f51e527071a service nova] Lock "21715aa5-24d4-423b-92a0-be7cc1f60877-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 929.975805] env[65522]: DEBUG oslo_concurrency.lockutils [req-0c35b74e-acd1-47ec-8c50-a1a3f518c59c req-cebc6420-b8ca-4407-81e3-5f51e527071a service nova] Lock "21715aa5-24d4-423b-92a0-be7cc1f60877-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.975981] env[65522]: DEBUG nova.compute.manager [req-0c35b74e-acd1-47ec-8c50-a1a3f518c59c req-cebc6420-b8ca-4407-81e3-5f51e527071a service nova] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] No waiting events found dispatching network-vif-plugged-7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 929.976164] env[65522]: WARNING nova.compute.manager [req-0c35b74e-acd1-47ec-8c50-a1a3f518c59c req-cebc6420-b8ca-4407-81e3-5f51e527071a service nova] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Received unexpected event network-vif-plugged-7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880 for instance with vm_state building and task_state spawning. [ 929.985398] env[65522]: WARNING openstack [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 929.985763] env[65522]: WARNING openstack [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 929.992877] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b0fcd894-9789-4556-8f56-b1923c3709d2 tempest-ServerGroupTestJSON-758797606 tempest-ServerGroupTestJSON-758797606-project-member] Lock "55ab4771-9908-4640-a142-3cb40c0c6ee5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.517s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 930.012100] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114337, 'name': ReconfigVM_Task, 'duration_secs': 0.828014} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.014966] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 4aca4e87-b923-49e4-88d8-7ff51bcbe993/4aca4e87-b923-49e4-88d8-7ff51bcbe993.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.017869] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e73f04fb-607c-41eb-a0b8-0d566d3c8959 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.023927] env[65522]: INFO nova.compute.manager [None req-3f5b81ae-016e-40f4-9508-1044fb7115c8 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance to original state: 'active' [ 930.028961] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 930.028961] env[65522]: value = "task-5114342" [ 930.028961] env[65522]: _type = "Task" [ 930.028961] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.044377] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114342, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.101440] env[65522]: WARNING neutronclient.v2_0.client [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 930.102201] env[65522]: WARNING openstack [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.102548] env[65522]: WARNING openstack [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.159719] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114340, 'name': CreateVM_Task, 'duration_secs': 0.747207} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.160661] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 930.160661] env[65522]: WARNING neutronclient.v2_0.client [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 930.161997] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.161997] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 930.161997] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 930.162385] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea429d12-cbc9-43f8-abb7-e4b2d037b89d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.168815] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Waiting for the task: (returnval){ [ 930.168815] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52921a1c-1069-fb31-feaa-0c86cb427795" [ 930.168815] env[65522]: _type = "Task" [ 930.168815] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.181400] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52921a1c-1069-fb31-feaa-0c86cb427795, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.244695] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 930.290124] env[65522]: DEBUG nova.network.neutron [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Updated VIF entry in instance network info cache for port 41b80f01-c268-48b3-8886-1273938ac6bd. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 930.290612] env[65522]: DEBUG nova.network.neutron [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Updating instance_info_cache with network_info: [{"id": "41b80f01-c268-48b3-8886-1273938ac6bd", "address": "fa:16:3e:7d:ff:88", "network": {"id": "08114cc1-8575-4999-ac80-0c6d101dc304", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-264300140-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5ea1965645e741f6912422c21417f1ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9875d38f-76e2-416c-bfb7-f18a22b0d8ee", "external-id": "nsx-vlan-transportzone-442", "segmentation_id": 442, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41b80f01-c2", "ovs_interfaceid": "41b80f01-c268-48b3-8886-1273938ac6bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 930.309523] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Acquiring lock "refresh_cache-21715aa5-24d4-423b-92a0-be7cc1f60877" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.309712] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Acquired lock "refresh_cache-21715aa5-24d4-423b-92a0-be7cc1f60877" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 930.309970] env[65522]: DEBUG nova.network.neutron [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 930.484039] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb66ec1e-3bbb-4a26-96be-b43ec3175ed7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.493508] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74353a2f-1903-442a-a79d-da94c6f6b6dd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.528370] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f500a6-3647-4e1a-9b49-0f7819897a55 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.545133] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae417b94-fcfd-4e9e-8036-fa8009d1f6ab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.553880] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114342, 'name': Rename_Task, 'duration_secs': 0.272611} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.554572] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 930.554901] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80a1ce14-be72-40c9-a58b-5d2277c0f132 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.565110] env[65522]: DEBUG nova.compute.provider_tree [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.572834] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 930.572834] env[65522]: value = "task-5114343" [ 930.572834] env[65522]: _type = "Task" [ 930.572834] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.582288] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114343, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.685258] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52921a1c-1069-fb31-feaa-0c86cb427795, 'name': SearchDatastore_Task, 'duration_secs': 0.018296} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.685847] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 930.686153] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 930.686486] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.686717] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 930.687060] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 930.687481] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41a91928-79cf-46fc-8907-2c6ca5ae59a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.701363] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 930.701634] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 930.702466] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a6db7c5-baf1-4356-8fee-2adefcb6ffa6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.709150] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Waiting for the task: (returnval){ [ 930.709150] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528552ee-f70f-9840-98ad-dc3d8663d4e4" [ 930.709150] env[65522]: _type = "Task" [ 930.709150] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.719200] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528552ee-f70f-9840-98ad-dc3d8663d4e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.796719] env[65522]: DEBUG oslo_concurrency.lockutils [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] Releasing lock "refresh_cache-2a12a0a3-913e-4ade-85ef-cca7ba508ac9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 930.797056] env[65522]: DEBUG nova.compute.manager [req-d88c7ece-fc76-49d8-8646-d13a5d19266b req-129576f6-d80e-4d69-b901-0fc1a7431bd7 service nova] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Received event network-vif-deleted-d4e8efaa-adbc-4eec-adf0-e3f651352ba9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 930.813822] env[65522]: WARNING openstack [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.814224] env[65522]: WARNING openstack [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.855237] env[65522]: DEBUG nova.network.neutron [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 930.875390] env[65522]: WARNING openstack [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.875499] env[65522]: WARNING openstack [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.959293] env[65522]: WARNING neutronclient.v2_0.client [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 930.959988] env[65522]: WARNING openstack [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.960965] env[65522]: WARNING openstack [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 931.069071] env[65522]: DEBUG nova.scheduler.client.report [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 931.084218] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114343, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.106941] env[65522]: DEBUG nova.network.neutron [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Updating instance_info_cache with network_info: [{"id": "7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880", "address": "fa:16:3e:5c:bc:6f", "network": {"id": "ec1ee640-e97d-4f29-9441-852ee8bd090c", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-567256478-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3c9f139a53f47d8bfcb22fa53d7b7b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f6ad0cf-a0", "ovs_interfaceid": "7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 931.223121] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528552ee-f70f-9840-98ad-dc3d8663d4e4, 'name': SearchDatastore_Task, 'duration_secs': 0.013996} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.224089] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf5c026d-d7f8-4be2-8240-7b6f298b58db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.226952] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 931.227237] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 931.227504] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 931.227763] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 931.227931] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 931.231027] env[65522]: INFO nova.compute.manager [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Terminating instance [ 931.236217] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Waiting for the task: (returnval){ [ 931.236217] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520e6cca-aef4-52a8-771b-e77b20267ac6" [ 931.236217] env[65522]: _type = "Task" [ 931.236217] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.248514] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520e6cca-aef4-52a8-771b-e77b20267ac6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.330196] env[65522]: DEBUG nova.compute.manager [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Received event network-changed-c2eab4eb-5a45-4c96-a145-bea8b314c82d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 931.330196] env[65522]: DEBUG nova.compute.manager [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Refreshing instance network info cache due to event network-changed-c2eab4eb-5a45-4c96-a145-bea8b314c82d. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 931.330196] env[65522]: DEBUG oslo_concurrency.lockutils [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] Acquiring lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.330196] env[65522]: DEBUG oslo_concurrency.lockutils [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] Acquired lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 931.330400] env[65522]: DEBUG nova.network.neutron [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Refreshing network info cache for port c2eab4eb-5a45-4c96-a145-bea8b314c82d {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 931.574760] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.140s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 931.581412] env[65522]: DEBUG oslo_concurrency.lockutils [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.625s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 931.581412] env[65522]: DEBUG nova.objects.instance [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lazy-loading 'resources' on Instance uuid 75c8848b-aea9-43f1-8697-9224050d1fef {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.591587] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114343, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.602558] env[65522]: INFO nova.scheduler.client.report [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleted allocations for instance 1e7b8237-34ea-479e-b5a7-b1846661d61d [ 931.610500] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Releasing lock "refresh_cache-21715aa5-24d4-423b-92a0-be7cc1f60877" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 931.610971] env[65522]: DEBUG nova.compute.manager [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Instance network_info: |[{"id": "7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880", "address": "fa:16:3e:5c:bc:6f", "network": {"id": "ec1ee640-e97d-4f29-9441-852ee8bd090c", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-567256478-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3c9f139a53f47d8bfcb22fa53d7b7b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f6ad0cf-a0", "ovs_interfaceid": "7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 931.611531] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:bc:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d0c6fd7-3cc9-4818-9475-8f15900394cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 931.621827] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Creating folder: Project (d3c9f139a53f47d8bfcb22fa53d7b7b6). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 931.622630] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d37a8a15-a2c2-4e9d-a53b-5a7ecd3b46cd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.636895] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Created folder: Project (d3c9f139a53f47d8bfcb22fa53d7b7b6) in parent group-v994660. [ 931.636895] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Creating folder: Instances. Parent ref: group-v994853. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 931.636895] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aca3f3f3-c1ce-4a05-abfe-822bd11726a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.650113] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Created folder: Instances in parent group-v994853. [ 931.650528] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 931.650759] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 931.650994] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fbc7f4d8-5473-4c26-aca7-319e29047297 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.673509] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 931.673509] env[65522]: value = "task-5114346" [ 931.673509] env[65522]: _type = "Task" [ 931.673509] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.683711] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114346, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.738049] env[65522]: DEBUG nova.compute.manager [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 931.738427] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 931.740409] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-364b27bc-6dc9-47b9-8aa3-bda6d795d5b7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.762271] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520e6cca-aef4-52a8-771b-e77b20267ac6, 'name': SearchDatastore_Task, 'duration_secs': 0.017296} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.766318] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 931.766842] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 2a12a0a3-913e-4ade-85ef-cca7ba508ac9/2a12a0a3-913e-4ade-85ef-cca7ba508ac9.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 931.767341] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 931.767965] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ee379dd7-3ddf-4b46-9cdd-6a5b81508615 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.774482] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b1086f9-a87f-4f1c-99d8-ace917187b05 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.785309] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Waiting for the task: (returnval){ [ 931.785309] env[65522]: value = "task-5114348" [ 931.785309] env[65522]: _type = "Task" [ 931.785309] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.787806] env[65522]: DEBUG oslo_vmware.api [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 931.787806] env[65522]: value = "task-5114347" [ 931.787806] env[65522]: _type = "Task" [ 931.787806] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.802624] env[65522]: DEBUG oslo_vmware.api [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114347, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.807346] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114348, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.833498] env[65522]: WARNING neutronclient.v2_0.client [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 931.834334] env[65522]: WARNING openstack [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 931.834757] env[65522]: WARNING openstack [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 932.075054] env[65522]: WARNING openstack [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 932.075398] env[65522]: WARNING openstack [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 932.097492] env[65522]: DEBUG oslo_vmware.api [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114343, 'name': PowerOnVM_Task, 'duration_secs': 1.236914} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.097857] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.097966] env[65522]: INFO nova.compute.manager [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Took 11.39 seconds to spawn the instance on the hypervisor. [ 932.098276] env[65522]: DEBUG nova.compute.manager [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 932.099130] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff2a8fd0-d4d9-4c8c-a823-92af8731e407 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.105205] env[65522]: DEBUG nova.compute.manager [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Received event network-changed-7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 932.105404] env[65522]: DEBUG nova.compute.manager [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Refreshing instance network info cache due to event network-changed-7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 932.105694] env[65522]: DEBUG oslo_concurrency.lockutils [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] Acquiring lock "refresh_cache-21715aa5-24d4-423b-92a0-be7cc1f60877" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.105861] env[65522]: DEBUG oslo_concurrency.lockutils [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] Acquired lock "refresh_cache-21715aa5-24d4-423b-92a0-be7cc1f60877" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 932.106568] env[65522]: DEBUG nova.network.neutron [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Refreshing network info cache for port 7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 932.130613] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d18bfb92-1ac0-48c2-b11e-088f044feba2 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "1e7b8237-34ea-479e-b5a7-b1846661d61d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.506s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 932.179014] env[65522]: WARNING neutronclient.v2_0.client [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 932.179743] env[65522]: WARNING openstack [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 932.180103] env[65522]: WARNING openstack [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 932.206068] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114346, 'name': CreateVM_Task, 'duration_secs': 0.439719} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.206068] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 932.206441] env[65522]: WARNING neutronclient.v2_0.client [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 932.206894] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.207101] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 932.207478] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 932.207840] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a7e5f58-7d41-4fbf-9eed-a52d06894caa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.219458] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Waiting for the task: (returnval){ [ 932.219458] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fa6f53-4211-79f1-ad49-e34aab5b4e95" [ 932.219458] env[65522]: _type = "Task" [ 932.219458] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.230473] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fa6f53-4211-79f1-ad49-e34aab5b4e95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.305616] env[65522]: DEBUG oslo_vmware.api [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114347, 'name': PowerOffVM_Task, 'duration_secs': 0.280168} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.309237] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 932.309492] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 932.309795] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114348, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.310050] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7849e6fb-23d7-49c5-a6ff-1f1cbc68c2a4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.399239] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 932.399680] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 932.399996] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Deleting the datastore file [datastore1] 5e227f6c-9f13-416f-8e6d-2f7d931619fd {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 932.400543] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d29391d-80cd-4a6f-9cf6-25a2e1a3c7f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.412965] env[65522]: DEBUG nova.network.neutron [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Updated VIF entry in instance network info cache for port c2eab4eb-5a45-4c96-a145-bea8b314c82d. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 932.412965] env[65522]: DEBUG nova.network.neutron [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Updating instance_info_cache with network_info: [{"id": "c2eab4eb-5a45-4c96-a145-bea8b314c82d", "address": "fa:16:3e:22:77:b4", "network": {"id": "f49e9b61-6927-4a5b-a12d-f54f408d42b4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1397158171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a80f979f3dc0477e9462b47f7aa87f14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2eab4eb-5a", "ovs_interfaceid": "c2eab4eb-5a45-4c96-a145-bea8b314c82d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 932.419487] env[65522]: DEBUG oslo_vmware.api [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 932.419487] env[65522]: value = "task-5114350" [ 932.419487] env[65522]: _type = "Task" [ 932.419487] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.435422] env[65522]: DEBUG oslo_vmware.api [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114350, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.630027] env[65522]: WARNING neutronclient.v2_0.client [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 932.630027] env[65522]: WARNING openstack [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 932.630027] env[65522]: WARNING openstack [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 932.643045] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2dc69ab-27b2-47cb-9c03-1bf725d66768 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.654428] env[65522]: INFO nova.compute.manager [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Took 21.27 seconds to build instance. [ 932.657260] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c476c10b-ded0-40a4-87d2-335a2596807e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.700687] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbbe1f39-d818-407f-a15d-ff5f8a20efaa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.713429] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02ed26b-89e9-4fce-a138-031fc4d42305 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.737907] env[65522]: DEBUG nova.compute.provider_tree [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.747241] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fa6f53-4211-79f1-ad49-e34aab5b4e95, 'name': SearchDatastore_Task, 'duration_secs': 0.066962} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.747241] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 932.747327] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 932.748031] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.748031] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 932.748031] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 932.748388] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2dff0376-f1f1-42e1-9917-56ebe1573cec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.762500] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 932.762806] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 932.763817] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ece8641c-1bf9-4643-acf1-40e83111e71c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.784500] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Waiting for the task: (returnval){ [ 932.784500] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526adf66-a446-8a15-13db-0d3bd8419dcb" [ 932.784500] env[65522]: _type = "Task" [ 932.784500] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.792355] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526adf66-a446-8a15-13db-0d3bd8419dcb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.806298] env[65522]: WARNING openstack [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 932.806791] env[65522]: WARNING openstack [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 932.814322] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114348, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.721562} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.814892] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 2a12a0a3-913e-4ade-85ef-cca7ba508ac9/2a12a0a3-913e-4ade-85ef-cca7ba508ac9.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 932.815059] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 932.815347] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-115f9881-2fb9-4aa0-aae5-80b1db2a379d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.824553] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Waiting for the task: (returnval){ [ 932.824553] env[65522]: value = "task-5114351" [ 932.824553] env[65522]: _type = "Task" [ 932.824553] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.839234] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114351, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.920666] env[65522]: DEBUG oslo_concurrency.lockutils [req-b7463295-cc0a-40c7-8a72-e54b683b45ea req-82023338-750a-461b-8c1f-cf4e3dc995b1 service nova] Releasing lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 932.930730] env[65522]: DEBUG oslo_vmware.api [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114350, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.953857] env[65522]: WARNING neutronclient.v2_0.client [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 932.954522] env[65522]: WARNING openstack [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 932.954866] env[65522]: WARNING openstack [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 933.048599] env[65522]: DEBUG nova.network.neutron [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Updated VIF entry in instance network info cache for port 7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 933.050164] env[65522]: DEBUG nova.network.neutron [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Updating instance_info_cache with network_info: [{"id": "7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880", "address": "fa:16:3e:5c:bc:6f", "network": {"id": "ec1ee640-e97d-4f29-9441-852ee8bd090c", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-567256478-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3c9f139a53f47d8bfcb22fa53d7b7b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d0c6fd7-3cc9-4818-9475-8f15900394cc", "external-id": "nsx-vlan-transportzone-317", "segmentation_id": 317, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f6ad0cf-a0", "ovs_interfaceid": "7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 933.051677] env[65522]: DEBUG oslo_concurrency.lockutils [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "0b48b747-149c-4163-bdd0-7d0b07ddacfd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 933.051907] env[65522]: DEBUG oslo_concurrency.lockutils [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "0b48b747-149c-4163-bdd0-7d0b07ddacfd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 933.052122] env[65522]: DEBUG oslo_concurrency.lockutils [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "0b48b747-149c-4163-bdd0-7d0b07ddacfd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 933.052300] env[65522]: DEBUG oslo_concurrency.lockutils [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "0b48b747-149c-4163-bdd0-7d0b07ddacfd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 933.055018] env[65522]: DEBUG oslo_concurrency.lockutils [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "0b48b747-149c-4163-bdd0-7d0b07ddacfd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 933.055018] env[65522]: INFO nova.compute.manager [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Terminating instance [ 933.164427] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7f90de2b-01f3-492d-854d-d74dcc23af3c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "4aca4e87-b923-49e4-88d8-7ff51bcbe993" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.799s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 933.240784] env[65522]: DEBUG nova.scheduler.client.report [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 933.297089] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526adf66-a446-8a15-13db-0d3bd8419dcb, 'name': SearchDatastore_Task, 'duration_secs': 0.017422} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.298532] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d661162b-db81-4540-af74-3181ee5f49f9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.307113] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Waiting for the task: (returnval){ [ 933.307113] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529fbbdc-2af6-389f-7f14-b533b099e7cf" [ 933.307113] env[65522]: _type = "Task" [ 933.307113] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.321402] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529fbbdc-2af6-389f-7f14-b533b099e7cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.338603] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114351, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.198371} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.338983] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 933.340282] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e16369-ebc6-441a-bb65-5f61ef18e171 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.381332] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 2a12a0a3-913e-4ade-85ef-cca7ba508ac9/2a12a0a3-913e-4ade-85ef-cca7ba508ac9.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.381852] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c52c9e2-381e-46d2-be72-f0666943606d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.414864] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Waiting for the task: (returnval){ [ 933.414864] env[65522]: value = "task-5114352" [ 933.414864] env[65522]: _type = "Task" [ 933.414864] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.428181] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114352, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.436287] env[65522]: DEBUG oslo_vmware.api [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114350, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.711873} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.436612] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 933.436773] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 933.436995] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 933.437283] env[65522]: INFO nova.compute.manager [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Took 1.70 seconds to destroy the instance on the hypervisor. [ 933.437816] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 933.437926] env[65522]: DEBUG nova.compute.manager [-] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 933.438043] env[65522]: DEBUG nova.network.neutron [-] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 933.438385] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 933.439036] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 933.439918] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 933.515903] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 933.555286] env[65522]: DEBUG oslo_concurrency.lockutils [req-f5a25979-1910-445b-bd9d-91edc1cbbad9 req-9a3fd2a7-a7c9-4f7e-93fd-74fd0c846c50 service nova] Releasing lock "refresh_cache-21715aa5-24d4-423b-92a0-be7cc1f60877" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 933.559427] env[65522]: DEBUG nova.compute.manager [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 933.559640] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.560688] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43f69250-4062-40ef-ab62-5ad22971a4b2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.569139] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.569427] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b98c30a1-873d-4ebd-a7c7-3bc686179a15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.576873] env[65522]: DEBUG oslo_vmware.api [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 933.576873] env[65522]: value = "task-5114353" [ 933.576873] env[65522]: _type = "Task" [ 933.576873] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.586714] env[65522]: DEBUG oslo_vmware.api [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114353, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.749228] env[65522]: DEBUG oslo_concurrency.lockutils [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.169s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 933.752345] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.872s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 933.754719] env[65522]: INFO nova.compute.claims [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 933.777381] env[65522]: INFO nova.scheduler.client.report [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Deleted allocations for instance 75c8848b-aea9-43f1-8697-9224050d1fef [ 933.822063] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529fbbdc-2af6-389f-7f14-b533b099e7cf, 'name': SearchDatastore_Task, 'duration_secs': 0.066129} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.822522] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 933.823098] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 21715aa5-24d4-423b-92a0-be7cc1f60877/21715aa5-24d4-423b-92a0-be7cc1f60877.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 933.823918] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-35ab3a60-f50c-49fc-a8ba-b87d74fd9976 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.834613] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Waiting for the task: (returnval){ [ 933.834613] env[65522]: value = "task-5114354" [ 933.834613] env[65522]: _type = "Task" [ 933.834613] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.847870] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114354, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.928906] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114352, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.087543] env[65522]: DEBUG oslo_vmware.api [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114353, 'name': PowerOffVM_Task, 'duration_secs': 0.318024} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.087828] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 934.088009] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 934.088365] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-461cb55c-5b89-4760-83ed-3b395b421aef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.164154] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 934.164772] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 934.165060] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleting the datastore file [datastore1] 0b48b747-149c-4163-bdd0-7d0b07ddacfd {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.165377] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f4ffbbe3-8334-4a38-91f5-46c2603babce {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.173414] env[65522]: DEBUG oslo_vmware.api [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 934.173414] env[65522]: value = "task-5114356" [ 934.173414] env[65522]: _type = "Task" [ 934.173414] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.186137] env[65522]: DEBUG oslo_vmware.api [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114356, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.285418] env[65522]: DEBUG oslo_concurrency.lockutils [None req-48295973-d869-41b1-965b-6180a884fed1 tempest-SecurityGroupsTestJSON-1089212024 tempest-SecurityGroupsTestJSON-1089212024-project-member] Lock "75c8848b-aea9-43f1-8697-9224050d1fef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.748s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 934.343167] env[65522]: DEBUG nova.network.neutron [-] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 934.348490] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114354, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.431065] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114352, 'name': ReconfigVM_Task, 'duration_secs': 0.605485} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.431448] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 2a12a0a3-913e-4ade-85ef-cca7ba508ac9/2a12a0a3-913e-4ade-85ef-cca7ba508ac9.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 934.432251] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a1ec1ec3-07f8-4a21-92e0-e5e87efc4547 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.442970] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Waiting for the task: (returnval){ [ 934.442970] env[65522]: value = "task-5114357" [ 934.442970] env[65522]: _type = "Task" [ 934.442970] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.455859] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114357, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.511396] env[65522]: DEBUG nova.compute.manager [req-2b6969bc-6ec3-402a-9002-3f9b3f763ec5 req-35446f8a-a5f0-4379-aa73-427aef8e9bf9 service nova] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Received event network-vif-deleted-3f79042c-23dd-4fc0-b3cf-1e2884edabc1 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 934.547524] env[65522]: DEBUG oslo_concurrency.lockutils [None req-59a92715-7f8e-454a-b0b6-53c439a5e540 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "interface-9df1a51a-2811-4486-a4c6-58d618f2ae7d-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.547836] env[65522]: DEBUG oslo_concurrency.lockutils [None req-59a92715-7f8e-454a-b0b6-53c439a5e540 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-9df1a51a-2811-4486-a4c6-58d618f2ae7d-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.548240] env[65522]: DEBUG nova.objects.instance [None req-59a92715-7f8e-454a-b0b6-53c439a5e540 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'flavor' on Instance uuid 9df1a51a-2811-4486-a4c6-58d618f2ae7d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.687229] env[65522]: DEBUG oslo_vmware.api [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114356, 'name': DeleteDatastoreFile_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.771075] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "1b2779f5-c6e8-4226-a819-0560d63bd7df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.771206] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "1b2779f5-c6e8-4226-a819-0560d63bd7df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.851482] env[65522]: INFO nova.compute.manager [-] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Took 1.41 seconds to deallocate network for instance. [ 934.851813] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114354, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.957629] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114357, 'name': Rename_Task, 'duration_secs': 0.191276} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.957934] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 934.958476] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9100ea2c-20a5-43f4-a79d-b108dd1da4b4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.967949] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Waiting for the task: (returnval){ [ 934.967949] env[65522]: value = "task-5114358" [ 934.967949] env[65522]: _type = "Task" [ 934.967949] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.979770] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114358, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.052280] env[65522]: WARNING neutronclient.v2_0.client [None req-59a92715-7f8e-454a-b0b6-53c439a5e540 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.052998] env[65522]: WARNING openstack [None req-59a92715-7f8e-454a-b0b6-53c439a5e540 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.053399] env[65522]: WARNING openstack [None req-59a92715-7f8e-454a-b0b6-53c439a5e540 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.058758] env[65522]: DEBUG nova.objects.instance [None req-59a92715-7f8e-454a-b0b6-53c439a5e540 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'pci_requests' on Instance uuid 9df1a51a-2811-4486-a4c6-58d618f2ae7d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.173156] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a859f057-824a-491b-85de-2cbfe1953442 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.191420] env[65522]: DEBUG oslo_vmware.api [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114356, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.5149} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.193053] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ff2e79-61db-4891-ae46-ec86a3388237 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.197412] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.197947] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.198359] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.198775] env[65522]: INFO nova.compute.manager [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Took 1.64 seconds to destroy the instance on the hypervisor. [ 935.199274] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 935.199902] env[65522]: DEBUG nova.compute.manager [-] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 935.200617] env[65522]: DEBUG nova.network.neutron [-] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 935.200617] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.201286] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.201932] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.246273] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eae0db3d-fc76-4d82-bc74-72d2f4c4b1c2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.256508] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.259613] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df417be-e22d-42b3-a4c5-23a57dc3cfba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.276486] env[65522]: DEBUG nova.compute.manager [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 935.281812] env[65522]: DEBUG nova.compute.provider_tree [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.349866] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114354, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.358024] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 935.478010] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114358, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.564576] env[65522]: DEBUG nova.objects.base [None req-59a92715-7f8e-454a-b0b6-53c439a5e540 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Object Instance<9df1a51a-2811-4486-a4c6-58d618f2ae7d> lazy-loaded attributes: flavor,pci_requests {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 935.564811] env[65522]: DEBUG nova.network.neutron [None req-59a92715-7f8e-454a-b0b6-53c439a5e540 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 935.565184] env[65522]: WARNING neutronclient.v2_0.client [None req-59a92715-7f8e-454a-b0b6-53c439a5e540 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.565535] env[65522]: WARNING neutronclient.v2_0.client [None req-59a92715-7f8e-454a-b0b6-53c439a5e540 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.566205] env[65522]: WARNING openstack [None req-59a92715-7f8e-454a-b0b6-53c439a5e540 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.566613] env[65522]: WARNING openstack [None req-59a92715-7f8e-454a-b0b6-53c439a5e540 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.788919] env[65522]: DEBUG nova.scheduler.client.report [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 935.810299] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 935.848188] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114354, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.852719} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.848455] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 21715aa5-24d4-423b-92a0-be7cc1f60877/21715aa5-24d4-423b-92a0-be7cc1f60877.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 935.848670] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 935.848929] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7a1dc5a7-0847-43cd-a550-9d0ebf734900 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.856914] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Waiting for the task: (returnval){ [ 935.856914] env[65522]: value = "task-5114359" [ 935.856914] env[65522]: _type = "Task" [ 935.856914] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.868409] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114359, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.920293] env[65522]: DEBUG oslo_concurrency.lockutils [None req-59a92715-7f8e-454a-b0b6-53c439a5e540 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-9df1a51a-2811-4486-a4c6-58d618f2ae7d-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.372s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 935.980162] env[65522]: DEBUG oslo_vmware.api [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114358, 'name': PowerOnVM_Task, 'duration_secs': 0.785319} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.980162] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 935.980162] env[65522]: INFO nova.compute.manager [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Took 9.49 seconds to spawn the instance on the hypervisor. [ 935.980162] env[65522]: DEBUG nova.compute.manager [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 935.980768] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9921426-274e-4bb9-bec9-482bec67ddf2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.077639] env[65522]: DEBUG nova.network.neutron [-] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 936.296393] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.544s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 936.296939] env[65522]: DEBUG nova.compute.manager [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 936.300385] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.776s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 936.301831] env[65522]: INFO nova.compute.claims [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 936.369350] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114359, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083475} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.369651] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 936.371342] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e76b58b-c3af-43a0-b8fd-ca6e947bc6c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.399796] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] 21715aa5-24d4-423b-92a0-be7cc1f60877/21715aa5-24d4-423b-92a0-be7cc1f60877.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.400184] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47667c76-4b4d-4d66-a8aa-fb7b7ace559c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.422374] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Waiting for the task: (returnval){ [ 936.422374] env[65522]: value = "task-5114360" [ 936.422374] env[65522]: _type = "Task" [ 936.422374] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.433986] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114360, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.503781] env[65522]: INFO nova.compute.manager [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Took 21.79 seconds to build instance. [ 936.580419] env[65522]: INFO nova.compute.manager [-] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Took 1.38 seconds to deallocate network for instance. [ 936.807215] env[65522]: DEBUG nova.compute.utils [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 936.812784] env[65522]: DEBUG nova.compute.manager [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Not allocating networking since 'none' was specified. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 936.839028] env[65522]: DEBUG nova.compute.manager [req-baf0dee0-14a5-4e1d-9c7b-226d1ffc2e09 req-e38ee1bc-464f-498a-b64f-9ff41000261c service nova] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Received event network-vif-deleted-6f040560-31cf-4beb-916d-44d19ed60c9e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 936.936599] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114360, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.005966] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b97dd288-5645-4474-ac57-e0a615c930e1 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Lock "2a12a0a3-913e-4ade-85ef-cca7ba508ac9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.301s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 937.087805] env[65522]: DEBUG oslo_concurrency.lockutils [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.419556] env[65522]: DEBUG nova.compute.manager [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 937.435232] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114360, 'name': ReconfigVM_Task, 'duration_secs': 0.598852} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.435896] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Reconfigured VM instance instance-00000048 to attach disk [datastore1] 21715aa5-24d4-423b-92a0-be7cc1f60877/21715aa5-24d4-423b-92a0-be7cc1f60877.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.436612] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a4c72fc8-f3d6-4458-9705-346bbba0ae98 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.444635] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Waiting for the task: (returnval){ [ 937.444635] env[65522]: value = "task-5114361" [ 937.444635] env[65522]: _type = "Task" [ 937.444635] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.459525] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114361, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.525130] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Acquiring lock "2a12a0a3-913e-4ade-85ef-cca7ba508ac9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.525405] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Lock "2a12a0a3-913e-4ade-85ef-cca7ba508ac9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.525619] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Acquiring lock "2a12a0a3-913e-4ade-85ef-cca7ba508ac9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.525797] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Lock "2a12a0a3-913e-4ade-85ef-cca7ba508ac9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.525963] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Lock "2a12a0a3-913e-4ade-85ef-cca7ba508ac9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 937.528538] env[65522]: INFO nova.compute.manager [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Terminating instance [ 937.634598] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "interface-9df1a51a-2811-4486-a4c6-58d618f2ae7d-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.634861] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-9df1a51a-2811-4486-a4c6-58d618f2ae7d-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.635236] env[65522]: DEBUG nova.objects.instance [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'flavor' on Instance uuid 9df1a51a-2811-4486-a4c6-58d618f2ae7d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.812218] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c94c23-5325-413a-be7f-2576cdece64d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.820463] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43165215-a792-4456-8006-57d8cc4250fd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.854465] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a4847d-a6b9-422f-aae3-efffa4e14314 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.863573] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d18e7e2-b0ca-45f8-8008-37dc6da26852 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.877644] env[65522]: DEBUG nova.compute.provider_tree [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.956234] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114361, 'name': Rename_Task, 'duration_secs': 0.225108} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.956234] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 937.956514] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b9088615-ba13-4b1c-9e6f-458b46364274 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.964760] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Waiting for the task: (returnval){ [ 937.964760] env[65522]: value = "task-5114362" [ 937.964760] env[65522]: _type = "Task" [ 937.964760] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.973699] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114362, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.033365] env[65522]: DEBUG nova.compute.manager [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 938.033670] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.034625] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95bce56a-9c17-47ff-83e6-52a036820c67 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.044208] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 938.044531] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6317623d-328f-436d-b31d-524974581cfe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.052721] env[65522]: DEBUG oslo_vmware.api [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Waiting for the task: (returnval){ [ 938.052721] env[65522]: value = "task-5114363" [ 938.052721] env[65522]: _type = "Task" [ 938.052721] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.063905] env[65522]: DEBUG oslo_vmware.api [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114363, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.142952] env[65522]: WARNING neutronclient.v2_0.client [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 938.143638] env[65522]: WARNING openstack [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 938.144050] env[65522]: WARNING openstack [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 938.304543] env[65522]: DEBUG nova.objects.instance [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'pci_requests' on Instance uuid 9df1a51a-2811-4486-a4c6-58d618f2ae7d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.381175] env[65522]: DEBUG nova.scheduler.client.report [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 938.425738] env[65522]: DEBUG nova.compute.manager [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 938.458864] env[65522]: DEBUG nova.virt.hardware [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 938.458864] env[65522]: DEBUG nova.virt.hardware [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 938.458864] env[65522]: DEBUG nova.virt.hardware [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 938.458864] env[65522]: DEBUG nova.virt.hardware [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 938.459109] env[65522]: DEBUG nova.virt.hardware [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 938.459311] env[65522]: DEBUG nova.virt.hardware [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 938.459538] env[65522]: DEBUG nova.virt.hardware [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.459734] env[65522]: DEBUG nova.virt.hardware [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 938.459911] env[65522]: DEBUG nova.virt.hardware [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 938.460107] env[65522]: DEBUG nova.virt.hardware [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 938.460318] env[65522]: DEBUG nova.virt.hardware [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 938.461343] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa821cf-641b-4c08-aebd-ba321e743850 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.481568] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c539d7d-9763-4a00-9ccd-ee4555df6c3c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.485852] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114362, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.498599] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.504886] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Creating folder: Project (e0ce529e1abb4719a933fce138cd6c48). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 938.505364] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-40a002c6-16b9-4a8a-80dd-c0ce62826ef4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.517565] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Created folder: Project (e0ce529e1abb4719a933fce138cd6c48) in parent group-v994660. [ 938.517789] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Creating folder: Instances. Parent ref: group-v994856. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 938.518077] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9ca5813b-77c0-4a58-8be1-5c959715b300 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.528743] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Created folder: Instances in parent group-v994856. [ 938.529026] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 938.529297] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 938.529476] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-beb8e89b-c0a6-46ae-8e44-19dc44396262 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.550949] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.550949] env[65522]: value = "task-5114366" [ 938.550949] env[65522]: _type = "Task" [ 938.550949] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.564436] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114366, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.568125] env[65522]: DEBUG oslo_vmware.api [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114363, 'name': PowerOffVM_Task, 'duration_secs': 0.26229} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.568510] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 938.568686] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 938.568963] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61799112-2d61-4ce7-b892-01ed11926920 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.647888] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 938.648185] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 938.648771] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Deleting the datastore file [datastore1] 2a12a0a3-913e-4ade-85ef-cca7ba508ac9 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.648860] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-923362fa-c606-4a71-8c5b-65a74d6a63b7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.658907] env[65522]: DEBUG oslo_vmware.api [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Waiting for the task: (returnval){ [ 938.658907] env[65522]: value = "task-5114368" [ 938.658907] env[65522]: _type = "Task" [ 938.658907] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.672541] env[65522]: DEBUG oslo_vmware.api [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114368, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.809416] env[65522]: DEBUG nova.objects.base [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Object Instance<9df1a51a-2811-4486-a4c6-58d618f2ae7d> lazy-loaded attributes: flavor,pci_requests {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 938.809416] env[65522]: DEBUG nova.network.neutron [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 938.809793] env[65522]: WARNING neutronclient.v2_0.client [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 938.814027] env[65522]: WARNING neutronclient.v2_0.client [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 938.814027] env[65522]: WARNING openstack [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 938.814027] env[65522]: WARNING openstack [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 938.863653] env[65522]: DEBUG nova.policy [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b491939798e4481fb433ffb81f366b25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '068c2387de8c406194d9b1762c7292a7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 938.887271] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 938.887805] env[65522]: DEBUG nova.compute.manager [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 938.890931] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.447s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 938.891499] env[65522]: DEBUG nova.objects.instance [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lazy-loading 'resources' on Instance uuid 4264998b-949d-4b01-8832-fb1df6214403 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.976853] env[65522]: DEBUG oslo_vmware.api [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114362, 'name': PowerOnVM_Task, 'duration_secs': 0.614746} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.977819] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 938.977819] env[65522]: INFO nova.compute.manager [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Took 9.73 seconds to spawn the instance on the hypervisor. [ 938.977819] env[65522]: DEBUG nova.compute.manager [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 938.978613] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71653c36-d799-4c0c-a048-61f16caca7cd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.061161] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114366, 'name': CreateVM_Task, 'duration_secs': 0.455303} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.061572] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 939.061824] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.061991] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 939.062359] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 939.062608] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ba880ba-185d-4c64-afaf-9a991a4da70e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.068916] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Waiting for the task: (returnval){ [ 939.068916] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522e708d-10dc-4556-4f82-d03507244280" [ 939.068916] env[65522]: _type = "Task" [ 939.068916] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.077367] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522e708d-10dc-4556-4f82-d03507244280, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.176961] env[65522]: DEBUG oslo_vmware.api [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Task: {'id': task-5114368, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.276284} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.177321] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.177585] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 939.177879] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 939.178136] env[65522]: INFO nova.compute.manager [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Took 1.14 seconds to destroy the instance on the hypervisor. [ 939.178529] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 939.178818] env[65522]: DEBUG nova.compute.manager [-] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 939.178983] env[65522]: DEBUG nova.network.neutron [-] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 939.179320] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 939.180198] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.180910] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.203791] env[65522]: DEBUG nova.network.neutron [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Successfully created port: d91694fd-ef8c-478e-8e0e-4e8ca56702a8 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 939.291583] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 939.394599] env[65522]: DEBUG nova.compute.utils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 939.398614] env[65522]: DEBUG nova.compute.manager [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 939.398806] env[65522]: DEBUG nova.network.neutron [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 939.399113] env[65522]: WARNING neutronclient.v2_0.client [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 939.399409] env[65522]: WARNING neutronclient.v2_0.client [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 939.399958] env[65522]: WARNING openstack [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.400572] env[65522]: WARNING openstack [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.499700] env[65522]: INFO nova.compute.manager [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Took 24.21 seconds to build instance. [ 939.582185] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522e708d-10dc-4556-4f82-d03507244280, 'name': SearchDatastore_Task, 'duration_secs': 0.032723} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.583163] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 939.583163] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.583949] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.583949] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 939.583949] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.584627] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7ec3c63d-3f68-4468-805f-da1cf41669ab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.605849] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.606059] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 939.607114] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a303125-8460-4be9-88f0-f63451a962c7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.613810] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Waiting for the task: (returnval){ [ 939.613810] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526706ba-924a-6f12-e28b-a45cc8d69163" [ 939.613810] env[65522]: _type = "Task" [ 939.613810] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.625630] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526706ba-924a-6f12-e28b-a45cc8d69163, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.825098] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968820b7-c338-40cd-a615-2ea5060ac65b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.835039] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0ccafe-cc2c-465c-8222-7569b98a39d6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.871423] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0239fd02-15f9-4cfe-add4-f5073e5b083a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.880588] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d50ea89-dd90-4398-9fea-43afafd85cd5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.902943] env[65522]: DEBUG nova.compute.provider_tree [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.911828] env[65522]: DEBUG nova.compute.manager [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 940.001548] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4f47316-c84d-40f9-9bd9-4c37a2f6f3b2 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Lock "21715aa5-24d4-423b-92a0-be7cc1f60877" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.719s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 940.128108] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526706ba-924a-6f12-e28b-a45cc8d69163, 'name': SearchDatastore_Task, 'duration_secs': 0.024994} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.129260] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c2117a2-89a7-4f4c-8a88-257d364bc56c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.131765] env[65522]: DEBUG nova.network.neutron [-] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 940.137132] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Waiting for the task: (returnval){ [ 940.137132] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dee5cf-1027-9e19-feeb-65fa45e1238a" [ 940.137132] env[65522]: _type = "Task" [ 940.137132] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.147763] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dee5cf-1027-9e19-feeb-65fa45e1238a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.786587] env[65522]: DEBUG nova.scheduler.client.report [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 940.802543] env[65522]: INFO nova.compute.manager [-] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Took 1.62 seconds to deallocate network for instance. [ 940.817525] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dee5cf-1027-9e19-feeb-65fa45e1238a, 'name': SearchDatastore_Task, 'duration_secs': 0.018296} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.817867] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 940.818079] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2/9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 940.818468] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-efbc6c7d-7539-4fa6-aaeb-43caef074d69 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.828506] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Waiting for the task: (returnval){ [ 940.828506] env[65522]: value = "task-5114369" [ 940.828506] env[65522]: _type = "Task" [ 940.828506] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.838188] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114369, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.078701] env[65522]: DEBUG nova.network.neutron [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Successfully updated port: d91694fd-ef8c-478e-8e0e-4e8ca56702a8 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 941.127600] env[65522]: DEBUG nova.policy [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f010fe7784043aaa3d4d44b921c438f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91597233ae9c44c094f4c32d90332fa6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 941.307132] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.414s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 941.308688] env[65522]: DEBUG nova.compute.manager [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 941.313023] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.857s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 941.313352] env[65522]: INFO nova.compute.claims [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 941.316798] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 941.345640] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114369, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.348268] env[65522]: DEBUG nova.virt.hardware [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 941.348268] env[65522]: DEBUG nova.virt.hardware [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 941.348430] env[65522]: DEBUG nova.virt.hardware [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 941.348926] env[65522]: DEBUG nova.virt.hardware [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 941.348926] env[65522]: DEBUG nova.virt.hardware [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 941.348926] env[65522]: DEBUG nova.virt.hardware [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 941.349096] env[65522]: DEBUG nova.virt.hardware [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.349198] env[65522]: DEBUG nova.virt.hardware [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 941.349305] env[65522]: DEBUG nova.virt.hardware [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 941.349743] env[65522]: DEBUG nova.virt.hardware [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 941.349743] env[65522]: DEBUG nova.virt.hardware [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 941.350621] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69915c5a-6b13-4a73-9ac6-20f3d2291283 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.354094] env[65522]: INFO nova.scheduler.client.report [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleted allocations for instance 4264998b-949d-4b01-8832-fb1df6214403 [ 941.361714] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f39696-0d10-4a8e-bc9b-2938c85870d4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.554071] env[65522]: DEBUG nova.network.neutron [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Successfully created port: dd06b43a-4102-4a7a-a82e-cf76b354ba13 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 941.582815] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.583110] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 941.583364] env[65522]: DEBUG nova.network.neutron [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 941.846561] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114369, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.853329] env[65522]: DEBUG nova.compute.manager [req-1d3a7242-c4f0-4328-aab2-c1ede20842e5 req-600452b7-0f68-4766-904e-6663f32c212c service nova] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Received event network-vif-deleted-41b80f01-c268-48b3-8886-1273938ac6bd {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 941.865745] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e92070-b1df-4718-a704-39eac1b9d4dd tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "4264998b-949d-4b01-8832-fb1df6214403" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.426s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 942.088075] env[65522]: WARNING openstack [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.088770] env[65522]: WARNING openstack [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.136980] env[65522]: WARNING nova.network.neutron [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] 1ea0ec37-ea40-4911-b6ad-54375253cc42 already exists in list: networks containing: ['1ea0ec37-ea40-4911-b6ad-54375253cc42']. ignoring it [ 942.194142] env[65522]: WARNING openstack [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.194574] env[65522]: WARNING openstack [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.295789] env[65522]: WARNING neutronclient.v2_0.client [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 942.296646] env[65522]: WARNING openstack [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.297090] env[65522]: WARNING openstack [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.344549] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114369, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.088596} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.344867] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2/9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 942.345110] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 942.345423] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-afdc4449-5f9e-45fb-8355-23242dda5b15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.355070] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Waiting for the task: (returnval){ [ 942.355070] env[65522]: value = "task-5114370" [ 942.355070] env[65522]: _type = "Task" [ 942.355070] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.370253] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114370, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.446385] env[65522]: DEBUG nova.compute.manager [req-1783d085-95bd-483b-8be5-44c5fa438b80 req-ca5710c3-bd4b-47f1-adeb-9a006cf8e9db service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Received event network-vif-plugged-d91694fd-ef8c-478e-8e0e-4e8ca56702a8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 942.446801] env[65522]: DEBUG oslo_concurrency.lockutils [req-1783d085-95bd-483b-8be5-44c5fa438b80 req-ca5710c3-bd4b-47f1-adeb-9a006cf8e9db service nova] Acquiring lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 942.447188] env[65522]: DEBUG oslo_concurrency.lockutils [req-1783d085-95bd-483b-8be5-44c5fa438b80 req-ca5710c3-bd4b-47f1-adeb-9a006cf8e9db service nova] Lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 942.448978] env[65522]: DEBUG oslo_concurrency.lockutils [req-1783d085-95bd-483b-8be5-44c5fa438b80 req-ca5710c3-bd4b-47f1-adeb-9a006cf8e9db service nova] Lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 942.449273] env[65522]: DEBUG nova.compute.manager [req-1783d085-95bd-483b-8be5-44c5fa438b80 req-ca5710c3-bd4b-47f1-adeb-9a006cf8e9db service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] No waiting events found dispatching network-vif-plugged-d91694fd-ef8c-478e-8e0e-4e8ca56702a8 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 942.449495] env[65522]: WARNING nova.compute.manager [req-1783d085-95bd-483b-8be5-44c5fa438b80 req-ca5710c3-bd4b-47f1-adeb-9a006cf8e9db service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Received unexpected event network-vif-plugged-d91694fd-ef8c-478e-8e0e-4e8ca56702a8 for instance with vm_state active and task_state None. [ 942.472812] env[65522]: WARNING openstack [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.473235] env[65522]: WARNING openstack [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.545953] env[65522]: WARNING neutronclient.v2_0.client [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 942.546608] env[65522]: WARNING openstack [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.547043] env[65522]: WARNING openstack [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.663113] env[65522]: DEBUG nova.network.neutron [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updating instance_info_cache with network_info: [{"id": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "address": "fa:16:3e:8e:34:0c", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6221fb0-8c", "ovs_interfaceid": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d91694fd-ef8c-478e-8e0e-4e8ca56702a8", "address": "fa:16:3e:5d:61:15", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd91694fd-ef", "ovs_interfaceid": "d91694fd-ef8c-478e-8e0e-4e8ca56702a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 942.735891] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a1655a-bcd7-4da9-8954-978c508717d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.743552] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd2b54d-51c9-4eec-be4e-40f236885bf6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.780752] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fabbabaa-833b-4338-902a-935b3933409f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.790732] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf52b6bc-6447-4e73-9b46-f5ca43c27000 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.806017] env[65522]: DEBUG nova.compute.provider_tree [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.868340] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114370, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080953} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.868654] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 942.869473] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6616a4-5a42-4507-850d-a64d179cf5d5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.892120] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2/9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.892546] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c5d6633-d224-4560-8d96-6f9210c2a52e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.913804] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Waiting for the task: (returnval){ [ 942.913804] env[65522]: value = "task-5114371" [ 942.913804] env[65522]: _type = "Task" [ 942.913804] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.923207] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114371, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.167878] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 943.168929] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.169110] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 943.170202] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fcb730-1493-4531-bf4f-35a8892610a7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.190940] env[65522]: DEBUG nova.virt.hardware [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 943.191240] env[65522]: DEBUG nova.virt.hardware [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 943.191431] env[65522]: DEBUG nova.virt.hardware [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 943.191647] env[65522]: DEBUG nova.virt.hardware [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 943.191797] env[65522]: DEBUG nova.virt.hardware [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 943.192154] env[65522]: DEBUG nova.virt.hardware [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 943.192259] env[65522]: DEBUG nova.virt.hardware [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 943.192464] env[65522]: DEBUG nova.virt.hardware [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 943.192693] env[65522]: DEBUG nova.virt.hardware [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 943.192876] env[65522]: DEBUG nova.virt.hardware [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 943.193063] env[65522]: DEBUG nova.virt.hardware [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 943.201428] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Reconfiguring VM to attach interface {{(pid=65522) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 943.202133] env[65522]: DEBUG nova.network.neutron [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Successfully updated port: dd06b43a-4102-4a7a-a82e-cf76b354ba13 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 943.203306] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b78ce1a-b9c7-4f0e-8d29-ad44649f6c75 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.229061] env[65522]: DEBUG oslo_vmware.api [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 943.229061] env[65522]: value = "task-5114372" [ 943.229061] env[65522]: _type = "Task" [ 943.229061] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.239013] env[65522]: DEBUG oslo_vmware.api [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114372, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.309632] env[65522]: DEBUG nova.scheduler.client.report [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 943.425430] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114371, 'name': ReconfigVM_Task, 'duration_secs': 0.509049} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.425752] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2/9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 943.426424] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-24628595-a79e-46f2-9d14-4b7d5d15dc65 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.433504] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Waiting for the task: (returnval){ [ 943.433504] env[65522]: value = "task-5114373" [ 943.433504] env[65522]: _type = "Task" [ 943.433504] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.444435] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114373, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.630761] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "82a7ad85-a061-4eec-bd3f-fc977532dfbf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 943.630991] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "82a7ad85-a061-4eec-bd3f-fc977532dfbf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 943.723095] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "refresh_cache-84d4c689-b437-41a5-bb62-d144a30c6219" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.723363] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "refresh_cache-84d4c689-b437-41a5-bb62-d144a30c6219" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 943.723598] env[65522]: DEBUG nova.network.neutron [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 943.741143] env[65522]: DEBUG oslo_vmware.api [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114372, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.814900] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 943.815453] env[65522]: DEBUG nova.compute.manager [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 943.818216] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 17.961s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 943.926447] env[65522]: DEBUG nova.compute.manager [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Received event network-vif-plugged-dd06b43a-4102-4a7a-a82e-cf76b354ba13 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 943.926770] env[65522]: DEBUG oslo_concurrency.lockutils [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] Acquiring lock "84d4c689-b437-41a5-bb62-d144a30c6219-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 943.926880] env[65522]: DEBUG oslo_concurrency.lockutils [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] Lock "84d4c689-b437-41a5-bb62-d144a30c6219-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 943.927211] env[65522]: DEBUG oslo_concurrency.lockutils [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] Lock "84d4c689-b437-41a5-bb62-d144a30c6219-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 943.928318] env[65522]: DEBUG nova.compute.manager [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] No waiting events found dispatching network-vif-plugged-dd06b43a-4102-4a7a-a82e-cf76b354ba13 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 943.928318] env[65522]: WARNING nova.compute.manager [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Received unexpected event network-vif-plugged-dd06b43a-4102-4a7a-a82e-cf76b354ba13 for instance with vm_state building and task_state spawning. [ 943.928318] env[65522]: DEBUG nova.compute.manager [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Received event network-changed-dd06b43a-4102-4a7a-a82e-cf76b354ba13 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 943.928318] env[65522]: DEBUG nova.compute.manager [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Refreshing instance network info cache due to event network-changed-dd06b43a-4102-4a7a-a82e-cf76b354ba13. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 943.928318] env[65522]: DEBUG oslo_concurrency.lockutils [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] Acquiring lock "refresh_cache-84d4c689-b437-41a5-bb62-d144a30c6219" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.947566] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114373, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.135449] env[65522]: DEBUG nova.compute.manager [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 944.227041] env[65522]: WARNING openstack [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.227333] env[65522]: WARNING openstack [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.244746] env[65522]: DEBUG oslo_vmware.api [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114372, 'name': ReconfigVM_Task, 'duration_secs': 0.6832} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.244871] env[65522]: WARNING neutronclient.v2_0.client [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.245040] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 944.245261] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Reconfigured VM to attach interface {{(pid=65522) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 944.275983] env[65522]: DEBUG nova.network.neutron [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 944.297407] env[65522]: WARNING openstack [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.297837] env[65522]: WARNING openstack [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.321416] env[65522]: DEBUG nova.compute.utils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 944.325478] env[65522]: INFO nova.compute.claims [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.329058] env[65522]: DEBUG nova.compute.manager [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 944.329564] env[65522]: DEBUG nova.network.neutron [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 944.329564] env[65522]: WARNING neutronclient.v2_0.client [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.329864] env[65522]: WARNING neutronclient.v2_0.client [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.330442] env[65522]: WARNING openstack [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.330858] env[65522]: WARNING openstack [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.342091] env[65522]: DEBUG nova.compute.manager [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 944.387470] env[65522]: DEBUG nova.policy [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f51006be1b2e401a8dd5504e0e481810', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ea0fab7de6c47eeaf3bdde413ae3901', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 944.390329] env[65522]: WARNING neutronclient.v2_0.client [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.390952] env[65522]: WARNING openstack [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.391364] env[65522]: WARNING openstack [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.400598] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Acquiring lock "21715aa5-24d4-423b-92a0-be7cc1f60877" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 944.400816] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Lock "21715aa5-24d4-423b-92a0-be7cc1f60877" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 944.401012] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Acquiring lock "21715aa5-24d4-423b-92a0-be7cc1f60877-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 944.401200] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Lock "21715aa5-24d4-423b-92a0-be7cc1f60877-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 944.401362] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Lock "21715aa5-24d4-423b-92a0-be7cc1f60877-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 944.403183] env[65522]: INFO nova.compute.manager [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Terminating instance [ 944.452288] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114373, 'name': Rename_Task, 'duration_secs': 0.534348} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.452493] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 944.452759] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0a870daf-7a7a-4d18-9a27-5f390fc20314 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.460702] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Waiting for the task: (returnval){ [ 944.460702] env[65522]: value = "task-5114374" [ 944.460702] env[65522]: _type = "Task" [ 944.460702] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.469792] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114374, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.478375] env[65522]: DEBUG nova.network.neutron [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Updating instance_info_cache with network_info: [{"id": "dd06b43a-4102-4a7a-a82e-cf76b354ba13", "address": "fa:16:3e:36:f7:5d", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd06b43a-41", "ovs_interfaceid": "dd06b43a-4102-4a7a-a82e-cf76b354ba13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 944.629871] env[65522]: DEBUG nova.compute.manager [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Received event network-changed-d91694fd-ef8c-478e-8e0e-4e8ca56702a8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 944.630474] env[65522]: DEBUG nova.compute.manager [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Refreshing instance network info cache due to event network-changed-d91694fd-ef8c-478e-8e0e-4e8ca56702a8. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 944.630774] env[65522]: DEBUG oslo_concurrency.lockutils [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] Acquiring lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.631043] env[65522]: DEBUG oslo_concurrency.lockutils [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] Acquired lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 944.631220] env[65522]: DEBUG nova.network.neutron [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Refreshing network info cache for port d91694fd-ef8c-478e-8e0e-4e8ca56702a8 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 944.659866] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 944.685218] env[65522]: DEBUG nova.network.neutron [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Successfully created port: 5c7d9e50-0f3f-477e-af8d-86a5fe295457 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 944.750761] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cad2f162-ce02-408b-a722-be6522342d2f tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-9df1a51a-2811-4486-a4c6-58d618f2ae7d-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.115s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 944.848732] env[65522]: INFO nova.compute.resource_tracker [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating resource usage from migration 43f083bf-03be-46c8-ba8b-81c6563136b6 [ 944.906461] env[65522]: DEBUG nova.compute.manager [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 944.906659] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.907697] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b25f314c-ce9f-44f2-976e-af7f99794c85 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.920757] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.921056] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d60058be-b38c-48c2-9d59-c9722d43c797 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.929063] env[65522]: DEBUG oslo_vmware.api [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Waiting for the task: (returnval){ [ 944.929063] env[65522]: value = "task-5114375" [ 944.929063] env[65522]: _type = "Task" [ 944.929063] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.938501] env[65522]: DEBUG oslo_vmware.api [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114375, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.972531] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114374, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.983555] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "refresh_cache-84d4c689-b437-41a5-bb62-d144a30c6219" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 944.983968] env[65522]: DEBUG nova.compute.manager [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Instance network_info: |[{"id": "dd06b43a-4102-4a7a-a82e-cf76b354ba13", "address": "fa:16:3e:36:f7:5d", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd06b43a-41", "ovs_interfaceid": "dd06b43a-4102-4a7a-a82e-cf76b354ba13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 944.984554] env[65522]: DEBUG oslo_concurrency.lockutils [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] Acquired lock "refresh_cache-84d4c689-b437-41a5-bb62-d144a30c6219" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 944.984747] env[65522]: DEBUG nova.network.neutron [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Refreshing network info cache for port dd06b43a-4102-4a7a-a82e-cf76b354ba13 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 944.985981] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:f7:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56136ef6-99d7-4562-9a9f-d66fec951c5c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd06b43a-4102-4a7a-a82e-cf76b354ba13', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 944.994024] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 944.995586] env[65522]: WARNING neutronclient.v2_0.client [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.996295] env[65522]: WARNING openstack [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.996586] env[65522]: WARNING openstack [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 945.004073] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 945.007344] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b28d4300-7541-47b5-8b86-eeadc01de16c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.030479] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.030479] env[65522]: value = "task-5114376" [ 945.030479] env[65522]: _type = "Task" [ 945.030479] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.045944] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114376, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.104539] env[65522]: WARNING openstack [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 945.105067] env[65522]: WARNING openstack [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 945.133910] env[65522]: WARNING neutronclient.v2_0.client [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 945.134795] env[65522]: WARNING openstack [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 945.135430] env[65522]: WARNING openstack [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 945.249723] env[65522]: WARNING neutronclient.v2_0.client [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 945.250516] env[65522]: WARNING openstack [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 945.250885] env[65522]: WARNING openstack [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 945.322380] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88284e1e-85db-46af-a98a-92ee9eb2c234 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.331360] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09beea85-78e3-42f4-b58e-5a2c291d4e0f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.366332] env[65522]: DEBUG nova.compute.manager [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 945.373638] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210291ec-d6a7-4a05-9686-a9503eb9fe6f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.382470] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e493707-c85a-47cc-bc4d-a4ca2f575c9f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.388518] env[65522]: WARNING openstack [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 945.388779] env[65522]: WARNING openstack [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 945.407922] env[65522]: DEBUG nova.compute.provider_tree [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.411357] env[65522]: DEBUG nova.virt.hardware [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 945.411582] env[65522]: DEBUG nova.virt.hardware [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 945.411731] env[65522]: DEBUG nova.virt.hardware [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 945.411907] env[65522]: DEBUG nova.virt.hardware [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 945.412060] env[65522]: DEBUG nova.virt.hardware [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 945.412207] env[65522]: DEBUG nova.virt.hardware [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 945.412411] env[65522]: DEBUG nova.virt.hardware [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 945.412566] env[65522]: DEBUG nova.virt.hardware [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 945.412749] env[65522]: DEBUG nova.virt.hardware [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 945.412911] env[65522]: DEBUG nova.virt.hardware [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 945.413092] env[65522]: DEBUG nova.virt.hardware [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 945.413961] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1976307-65b4-468e-ac53-85ab500c2cce {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.417712] env[65522]: DEBUG nova.scheduler.client.report [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 945.427150] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0dd76fd-8ff5-4a35-9d85-565bb41bf2fa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.441419] env[65522]: DEBUG oslo_vmware.api [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114375, 'name': PowerOffVM_Task, 'duration_secs': 0.400366} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.449203] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 945.449456] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 945.450276] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eaa7972d-d200-4bcf-97fd-027430a98736 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.472828] env[65522]: DEBUG oslo_vmware.api [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114374, 'name': PowerOnVM_Task, 'duration_secs': 0.699997} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.473128] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 945.473758] env[65522]: INFO nova.compute.manager [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Took 7.05 seconds to spawn the instance on the hypervisor. [ 945.473758] env[65522]: DEBUG nova.compute.manager [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 945.474347] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23bf2e4-315c-44f0-8aa7-8eb271af2b9e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.488561] env[65522]: DEBUG nova.network.neutron [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Updated VIF entry in instance network info cache for port dd06b43a-4102-4a7a-a82e-cf76b354ba13. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 945.488894] env[65522]: DEBUG nova.network.neutron [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Updating instance_info_cache with network_info: [{"id": "dd06b43a-4102-4a7a-a82e-cf76b354ba13", "address": "fa:16:3e:36:f7:5d", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd06b43a-41", "ovs_interfaceid": "dd06b43a-4102-4a7a-a82e-cf76b354ba13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 945.527825] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 945.527825] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 945.527825] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Deleting the datastore file [datastore1] 21715aa5-24d4-423b-92a0-be7cc1f60877 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.528031] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0b4ac6fe-3283-4fbc-b82a-2d221a4b2ec0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.542362] env[65522]: DEBUG oslo_vmware.api [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Waiting for the task: (returnval){ [ 945.542362] env[65522]: value = "task-5114378" [ 945.542362] env[65522]: _type = "Task" [ 945.542362] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.554382] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114376, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.559833] env[65522]: DEBUG oslo_vmware.api [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114378, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.609373] env[65522]: WARNING neutronclient.v2_0.client [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 945.610101] env[65522]: WARNING openstack [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 945.610568] env[65522]: WARNING openstack [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 945.793763] env[65522]: DEBUG nova.network.neutron [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updated VIF entry in instance network info cache for port d91694fd-ef8c-478e-8e0e-4e8ca56702a8. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 945.794081] env[65522]: DEBUG nova.network.neutron [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updating instance_info_cache with network_info: [{"id": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "address": "fa:16:3e:8e:34:0c", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6221fb0-8c", "ovs_interfaceid": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d91694fd-ef8c-478e-8e0e-4e8ca56702a8", "address": "fa:16:3e:5d:61:15", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd91694fd-ef", "ovs_interfaceid": "d91694fd-ef8c-478e-8e0e-4e8ca56702a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 945.922446] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.104s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 945.922648] env[65522]: INFO nova.compute.manager [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Migrating [ 945.929337] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.149s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 945.930857] env[65522]: INFO nova.compute.claims [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 945.991821] env[65522]: DEBUG oslo_concurrency.lockutils [req-3e55f228-782d-48bc-a532-93993ce138f4 req-8677e092-bc54-4bbe-a30a-c6f8eecb1e7d service nova] Releasing lock "refresh_cache-84d4c689-b437-41a5-bb62-d144a30c6219" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 945.998626] env[65522]: INFO nova.compute.manager [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Took 23.15 seconds to build instance. [ 946.045201] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114376, 'name': CreateVM_Task, 'duration_secs': 0.5649} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.048288] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 946.048822] env[65522]: WARNING neutronclient.v2_0.client [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 946.049208] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.049360] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 946.049663] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 946.050428] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6826d5aa-44c8-4019-a039-c5c9d5ca1cc9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.056081] env[65522]: DEBUG oslo_vmware.api [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114378, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.057521] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 946.057521] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e9ef5f-e565-a3c9-f32b-0e324e90a21f" [ 946.057521] env[65522]: _type = "Task" [ 946.057521] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.066711] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e9ef5f-e565-a3c9-f32b-0e324e90a21f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.254630] env[65522]: DEBUG nova.network.neutron [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Successfully updated port: 5c7d9e50-0f3f-477e-af8d-86a5fe295457 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 946.296912] env[65522]: DEBUG oslo_concurrency.lockutils [req-1507d8b2-7ae0-4e72-851f-3d7f60a1cb11 req-712f68d2-9047-42cb-bc0c-6f1a2b553d50 service nova] Releasing lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 946.338120] env[65522]: DEBUG nova.compute.manager [req-0485c1be-7618-467c-95e8-50a36d486359 req-afa8ecf7-5281-4c88-9930-5b6e520bf1ed service nova] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Received event network-vif-plugged-5c7d9e50-0f3f-477e-af8d-86a5fe295457 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 946.338390] env[65522]: DEBUG oslo_concurrency.lockutils [req-0485c1be-7618-467c-95e8-50a36d486359 req-afa8ecf7-5281-4c88-9930-5b6e520bf1ed service nova] Acquiring lock "8da2a500-6f0e-4eda-9a92-79510753b3b8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 946.338551] env[65522]: DEBUG oslo_concurrency.lockutils [req-0485c1be-7618-467c-95e8-50a36d486359 req-afa8ecf7-5281-4c88-9930-5b6e520bf1ed service nova] Lock "8da2a500-6f0e-4eda-9a92-79510753b3b8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 946.338725] env[65522]: DEBUG oslo_concurrency.lockutils [req-0485c1be-7618-467c-95e8-50a36d486359 req-afa8ecf7-5281-4c88-9930-5b6e520bf1ed service nova] Lock "8da2a500-6f0e-4eda-9a92-79510753b3b8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 946.338931] env[65522]: DEBUG nova.compute.manager [req-0485c1be-7618-467c-95e8-50a36d486359 req-afa8ecf7-5281-4c88-9930-5b6e520bf1ed service nova] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] No waiting events found dispatching network-vif-plugged-5c7d9e50-0f3f-477e-af8d-86a5fe295457 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 946.339073] env[65522]: WARNING nova.compute.manager [req-0485c1be-7618-467c-95e8-50a36d486359 req-afa8ecf7-5281-4c88-9930-5b6e520bf1ed service nova] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Received unexpected event network-vif-plugged-5c7d9e50-0f3f-477e-af8d-86a5fe295457 for instance with vm_state building and task_state spawning. [ 946.444106] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.444350] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 946.444537] env[65522]: DEBUG nova.network.neutron [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 946.501767] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e0d9e68a-5a75-4468-b82a-c7b3d712c017 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Lock "9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.662s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 946.558353] env[65522]: DEBUG oslo_vmware.api [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Task: {'id': task-5114378, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.646398} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.563196] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.563498] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 946.563758] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 946.564031] env[65522]: INFO nova.compute.manager [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Took 1.66 seconds to destroy the instance on the hypervisor. [ 946.564434] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 946.565224] env[65522]: DEBUG nova.compute.manager [-] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 946.565373] env[65522]: DEBUG nova.network.neutron [-] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 946.565757] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 946.566587] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 946.566991] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 946.584363] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e9ef5f-e565-a3c9-f32b-0e324e90a21f, 'name': SearchDatastore_Task, 'duration_secs': 0.013518} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.584769] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 946.585549] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.585549] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.585549] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 946.585692] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.585848] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-166d9a39-963f-4484-b9f7-c2a5ad5c5e59 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.595878] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.596108] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 946.596858] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-515ef4f6-7fc3-4b13-b86e-8a38eec4c111 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.603735] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 946.603735] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525f3fdb-08a5-d2c1-b0bd-fe51a5bb51e6" [ 946.603735] env[65522]: _type = "Task" [ 946.603735] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.610501] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 946.619872] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525f3fdb-08a5-d2c1-b0bd-fe51a5bb51e6, 'name': SearchDatastore_Task, 'duration_secs': 0.013139} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.620798] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92b1ae3c-3d09-464c-b48f-9829fd046dd5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.627313] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 946.627313] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ec8800-8883-86ca-eba5-27be65707103" [ 946.627313] env[65522]: _type = "Task" [ 946.627313] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.636966] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ec8800-8883-86ca-eba5-27be65707103, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.757239] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "refresh_cache-8da2a500-6f0e-4eda-9a92-79510753b3b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.757376] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "refresh_cache-8da2a500-6f0e-4eda-9a92-79510753b3b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 946.757627] env[65522]: DEBUG nova.network.neutron [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 946.950376] env[65522]: WARNING neutronclient.v2_0.client [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 946.950698] env[65522]: WARNING openstack [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 946.951351] env[65522]: WARNING openstack [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 946.996244] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "interface-9df1a51a-2811-4486-a4c6-58d618f2ae7d-04e1485f-54a3-42d2-bd3f-19ea0fa7babe" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 946.996513] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-9df1a51a-2811-4486-a4c6-58d618f2ae7d-04e1485f-54a3-42d2-bd3f-19ea0fa7babe" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 946.996921] env[65522]: DEBUG nova.objects.instance [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'flavor' on Instance uuid 9df1a51a-2811-4486-a4c6-58d618f2ae7d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.073793] env[65522]: WARNING openstack [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.074203] env[65522]: WARNING openstack [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.141629] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ec8800-8883-86ca-eba5-27be65707103, 'name': SearchDatastore_Task, 'duration_secs': 0.012505} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.142913] env[65522]: WARNING neutronclient.v2_0.client [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.144513] env[65522]: WARNING openstack [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.144846] env[65522]: WARNING openstack [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.154671] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 947.154931] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 84d4c689-b437-41a5-bb62-d144a30c6219/84d4c689-b437-41a5-bb62-d144a30c6219.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 947.155806] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88f67f2d-1556-45ec-be23-28c844c4fd4e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.164633] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 947.164633] env[65522]: value = "task-5114379" [ 947.164633] env[65522]: _type = "Task" [ 947.164633] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.177299] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114379, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.260402] env[65522]: WARNING openstack [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.261821] env[65522]: WARNING openstack [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.271569] env[65522]: DEBUG nova.network.neutron [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating instance_info_cache with network_info: [{"id": "f8145f30-4e64-4a22-89a7-ffa393891ffb", "address": "fa:16:3e:42:70:f5", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8145f30-4e", "ovs_interfaceid": "f8145f30-4e64-4a22-89a7-ffa393891ffb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 947.316311] env[65522]: DEBUG nova.network.neutron [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 947.318634] env[65522]: DEBUG nova.network.neutron [-] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 947.344197] env[65522]: WARNING openstack [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.344588] env[65522]: WARNING openstack [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.388914] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36f2a99-2007-4454-97dd-af42edb755ca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.402832] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-876d53df-700e-4086-8468-bd4d8cb9c802 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.448372] env[65522]: DEBUG nova.compute.manager [None req-67c72df7-906b-4b1a-99c1-c45f843f53f6 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 947.449937] env[65522]: WARNING neutronclient.v2_0.client [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.450664] env[65522]: WARNING openstack [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.451122] env[65522]: WARNING openstack [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.460352] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800a9043-f930-4a96-b4b7-36ea7c355873 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.464810] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d014ca-1c17-430f-b49f-8114f5393885 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.480886] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33adb479-6513-4c2e-af7a-2129d8a10a34 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.502013] env[65522]: DEBUG nova.compute.provider_tree [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.504998] env[65522]: WARNING neutronclient.v2_0.client [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.505558] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.505899] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.583534] env[65522]: DEBUG nova.network.neutron [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Updating instance_info_cache with network_info: [{"id": "5c7d9e50-0f3f-477e-af8d-86a5fe295457", "address": "fa:16:3e:cf:68:35", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c7d9e50-0f", "ovs_interfaceid": "5c7d9e50-0f3f-477e-af8d-86a5fe295457", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 947.648840] env[65522]: DEBUG nova.objects.instance [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'pci_requests' on Instance uuid 9df1a51a-2811-4486-a4c6-58d618f2ae7d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.677637] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114379, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.777372] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 947.821324] env[65522]: INFO nova.compute.manager [-] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Took 1.26 seconds to deallocate network for instance. [ 947.898448] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "05319f69-2ac7-4649-a90a-3542d87a3e82" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 947.898894] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "05319f69-2ac7-4649-a90a-3542d87a3e82" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 947.899276] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "05319f69-2ac7-4649-a90a-3542d87a3e82-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 947.899638] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "05319f69-2ac7-4649-a90a-3542d87a3e82-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 947.899905] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "05319f69-2ac7-4649-a90a-3542d87a3e82-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 947.902640] env[65522]: INFO nova.compute.manager [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Terminating instance [ 947.980488] env[65522]: INFO nova.compute.manager [None req-67c72df7-906b-4b1a-99c1-c45f843f53f6 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] instance snapshotting [ 947.981151] env[65522]: DEBUG nova.objects.instance [None req-67c72df7-906b-4b1a-99c1-c45f843f53f6 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Lazy-loading 'flavor' on Instance uuid 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.004507] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Acquiring lock "9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 948.004722] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Lock "9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 948.004963] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Acquiring lock "9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 948.005167] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Lock "9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 948.005334] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Lock "9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 948.007502] env[65522]: INFO nova.compute.manager [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Terminating instance [ 948.015708] env[65522]: DEBUG nova.scheduler.client.report [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 948.086995] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "refresh_cache-8da2a500-6f0e-4eda-9a92-79510753b3b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 948.087405] env[65522]: DEBUG nova.compute.manager [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Instance network_info: |[{"id": "5c7d9e50-0f3f-477e-af8d-86a5fe295457", "address": "fa:16:3e:cf:68:35", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c7d9e50-0f", "ovs_interfaceid": "5c7d9e50-0f3f-477e-af8d-86a5fe295457", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 948.087879] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:68:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5e1049e8-c06b-4c93-a9e1-2cbb530f3f95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c7d9e50-0f3f-477e-af8d-86a5fe295457', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 948.096344] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 948.096598] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 948.096833] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e93db829-f3d0-4c8e-9913-95daee5945ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.121321] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 948.121321] env[65522]: value = "task-5114380" [ 948.121321] env[65522]: _type = "Task" [ 948.121321] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.131578] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114380, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.152064] env[65522]: DEBUG nova.objects.base [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Object Instance<9df1a51a-2811-4486-a4c6-58d618f2ae7d> lazy-loaded attributes: flavor,pci_requests {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 948.152354] env[65522]: DEBUG nova.network.neutron [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 948.152808] env[65522]: WARNING neutronclient.v2_0.client [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 948.153242] env[65522]: WARNING neutronclient.v2_0.client [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 948.154102] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.154671] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.178297] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114379, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.611638} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.179117] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 84d4c689-b437-41a5-bb62-d144a30c6219/84d4c689-b437-41a5-bb62-d144a30c6219.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 948.179409] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.179753] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d10f72c4-5cbe-44dc-8ae4-1526fae76b43 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.188298] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 948.188298] env[65522]: value = "task-5114381" [ 948.188298] env[65522]: _type = "Task" [ 948.188298] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.199640] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114381, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.292203] env[65522]: DEBUG nova.policy [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b491939798e4481fb433ffb81f366b25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '068c2387de8c406194d9b1762c7292a7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 948.328171] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 948.409183] env[65522]: DEBUG nova.compute.manager [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 948.409513] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 948.411130] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4330b6d0-ae7e-47bc-a87a-4c7467816c14 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.421020] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 948.421497] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50d4f40a-98c0-449c-b1cf-a3dee66b6ffe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.430619] env[65522]: DEBUG oslo_vmware.api [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 948.430619] env[65522]: value = "task-5114382" [ 948.430619] env[65522]: _type = "Task" [ 948.430619] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.440173] env[65522]: DEBUG oslo_vmware.api [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114382, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.487689] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f486d5-3484-4f28-b654-335f238aef34 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.507981] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bfd3e43-bd62-470f-bd30-35d1aa62ea31 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.511349] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Acquiring lock "refresh_cache-9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.511533] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Acquired lock "refresh_cache-9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 948.511708] env[65522]: DEBUG nova.network.neutron [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 948.523012] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.593s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 948.523012] env[65522]: DEBUG nova.compute.manager [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 948.526219] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.282s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 948.526439] env[65522]: DEBUG nova.objects.instance [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Lazy-loading 'resources' on Instance uuid 42a08cfb-d865-4967-a086-370a2ca98b7d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.568446] env[65522]: DEBUG nova.compute.manager [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Received event network-changed-5c7d9e50-0f3f-477e-af8d-86a5fe295457 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 948.569401] env[65522]: DEBUG nova.compute.manager [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Refreshing instance network info cache due to event network-changed-5c7d9e50-0f3f-477e-af8d-86a5fe295457. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 948.569798] env[65522]: DEBUG oslo_concurrency.lockutils [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] Acquiring lock "refresh_cache-8da2a500-6f0e-4eda-9a92-79510753b3b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.570121] env[65522]: DEBUG oslo_concurrency.lockutils [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] Acquired lock "refresh_cache-8da2a500-6f0e-4eda-9a92-79510753b3b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 948.573423] env[65522]: DEBUG nova.network.neutron [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Refreshing network info cache for port 5c7d9e50-0f3f-477e-af8d-86a5fe295457 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 948.632079] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114380, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.702299] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114381, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080173} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.702725] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.704633] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-263d673a-0b16-43a6-83ec-de3cf55ba9ac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.732762] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 84d4c689-b437-41a5-bb62-d144a30c6219/84d4c689-b437-41a5-bb62-d144a30c6219.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.733163] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6481cb24-97c7-4d91-b15a-0a61407c0089 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.755111] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 948.755111] env[65522]: value = "task-5114383" [ 948.755111] env[65522]: _type = "Task" [ 948.755111] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.769537] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114383, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.941835] env[65522]: DEBUG oslo_vmware.api [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114382, 'name': PowerOffVM_Task, 'duration_secs': 0.373813} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.942107] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 948.942277] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 948.942531] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4341ad69-f14c-4455-8625-6e1d72846102 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.015268] env[65522]: WARNING neutronclient.v2_0.client [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.016209] env[65522]: WARNING openstack [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.016747] env[65522]: WARNING openstack [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.026833] env[65522]: DEBUG nova.compute.manager [None req-67c72df7-906b-4b1a-99c1-c45f843f53f6 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Instance disappeared during snapshot {{(pid=65522) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4652}} [ 949.029084] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 949.029221] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 949.029961] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Deleting the datastore file [datastore1] 05319f69-2ac7-4649-a90a-3542d87a3e82 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.031153] env[65522]: DEBUG nova.compute.utils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 949.035503] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74e14251-a2da-4932-abf4-0d4ba78e7c99 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.038045] env[65522]: DEBUG nova.compute.manager [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 949.038266] env[65522]: DEBUG nova.network.neutron [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 949.038612] env[65522]: WARNING neutronclient.v2_0.client [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.039274] env[65522]: WARNING neutronclient.v2_0.client [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.039489] env[65522]: WARNING openstack [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.040428] env[65522]: WARNING openstack [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.052456] env[65522]: DEBUG oslo_vmware.api [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for the task: (returnval){ [ 949.052456] env[65522]: value = "task-5114385" [ 949.052456] env[65522]: _type = "Task" [ 949.052456] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.057640] env[65522]: DEBUG nova.network.neutron [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 949.066308] env[65522]: DEBUG oslo_vmware.api [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114385, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.075868] env[65522]: WARNING neutronclient.v2_0.client [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.076289] env[65522]: WARNING openstack [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.076678] env[65522]: WARNING openstack [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.110969] env[65522]: DEBUG nova.policy [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a074a1efea64590b1a596a9a081d158', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1546bbdca07c443d84abea1155cfde37', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 949.134555] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114380, 'name': CreateVM_Task, 'duration_secs': 0.853474} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.137539] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 949.139100] env[65522]: WARNING neutronclient.v2_0.client [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.139366] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.139767] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 949.140568] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 949.140568] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f081c73b-189f-47a6-a98b-189e556a38f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.147239] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 949.147239] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dd8e71-7c5a-e5c9-efca-fd6669f13bf6" [ 949.147239] env[65522]: _type = "Task" [ 949.147239] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.161181] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dd8e71-7c5a-e5c9-efca-fd6669f13bf6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.219551] env[65522]: DEBUG nova.compute.manager [None req-67c72df7-906b-4b1a-99c1-c45f843f53f6 tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Found 0 images (rotation: 2) {{(pid=65522) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 949.271785] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114383, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.299589] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f90d0df-494e-4636-b247-cfd28b151d05 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.322382] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating instance 'a2a5fb2d-62e0-4809-a01f-f1df66dca58f' progress to 0 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 949.353454] env[65522]: DEBUG nova.network.neutron [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 949.489478] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf1d2c6-a02d-4154-86e6-daa424ba61e2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.498050] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e02e129-8eb7-4e41-86f5-e19cf44b87c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.529901] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59eed77c-dfcb-4c90-be7b-bc43e0bee0d7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.538491] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242c4774-5b1d-479e-9bcd-24a050f3672a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.542701] env[65522]: DEBUG nova.compute.manager [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 949.555581] env[65522]: DEBUG nova.compute.provider_tree [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.566045] env[65522]: DEBUG oslo_vmware.api [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114385, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.660886] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dd8e71-7c5a-e5c9-efca-fd6669f13bf6, 'name': SearchDatastore_Task, 'duration_secs': 0.212559} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.661218] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 949.661567] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 949.661752] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.661897] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 949.662091] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 949.662372] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a55fd9c0-5ad3-48f0-9c42-7c5fe13ab213 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.674615] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 949.674865] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 949.675640] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8666348e-90a4-4e7f-8e70-9acc9a53439b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.682596] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 949.682596] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526f7ed8-f636-57fb-6565-ca52ac7c1e3c" [ 949.682596] env[65522]: _type = "Task" [ 949.682596] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.692183] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526f7ed8-f636-57fb-6565-ca52ac7c1e3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.693455] env[65522]: DEBUG nova.network.neutron [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Successfully created port: bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 949.748044] env[65522]: WARNING openstack [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.748044] env[65522]: WARNING openstack [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.766087] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114383, 'name': ReconfigVM_Task, 'duration_secs': 0.621341} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.766383] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 84d4c689-b437-41a5-bb62-d144a30c6219/84d4c689-b437-41a5-bb62-d144a30c6219.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.768046] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-caa6e24b-dfcb-4d15-ba9c-1952e0a8c459 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.778020] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 949.778020] env[65522]: value = "task-5114386" [ 949.778020] env[65522]: _type = "Task" [ 949.778020] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.786780] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114386, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.835032] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 949.835032] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1aa92394-6d3b-4bb0-8a8f-f84a727e0171 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.843369] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 949.843369] env[65522]: value = "task-5114387" [ 949.843369] env[65522]: _type = "Task" [ 949.843369] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.858154] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Releasing lock "refresh_cache-9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 949.859047] env[65522]: DEBUG nova.compute.manager [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 949.859508] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 949.860085] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114387, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.861468] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07979415-a9a3-41e5-bf6b-871877ddc3f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.872284] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 949.872779] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb710952-0e77-49c3-93ff-4fb9daeaa522 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.884547] env[65522]: DEBUG oslo_vmware.api [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Waiting for the task: (returnval){ [ 949.884547] env[65522]: value = "task-5114388" [ 949.884547] env[65522]: _type = "Task" [ 949.884547] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.892100] env[65522]: DEBUG nova.network.neutron [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Successfully updated port: 04e1485f-54a3-42d2-bd3f-19ea0fa7babe {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 949.902042] env[65522]: DEBUG oslo_vmware.api [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114388, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.969458] env[65522]: WARNING neutronclient.v2_0.client [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.970398] env[65522]: WARNING openstack [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.970984] env[65522]: WARNING openstack [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.066544] env[65522]: DEBUG nova.scheduler.client.report [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 950.069914] env[65522]: DEBUG oslo_vmware.api [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114385, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.080435] env[65522]: DEBUG nova.network.neutron [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Updated VIF entry in instance network info cache for port 5c7d9e50-0f3f-477e-af8d-86a5fe295457. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 950.080813] env[65522]: DEBUG nova.network.neutron [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Updating instance_info_cache with network_info: [{"id": "5c7d9e50-0f3f-477e-af8d-86a5fe295457", "address": "fa:16:3e:cf:68:35", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c7d9e50-0f", "ovs_interfaceid": "5c7d9e50-0f3f-477e-af8d-86a5fe295457", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 950.201642] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526f7ed8-f636-57fb-6565-ca52ac7c1e3c, 'name': SearchDatastore_Task, 'duration_secs': 0.218968} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.201642] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da1bf6df-5218-4c74-9931-274fe9ef08e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.206974] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 950.206974] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a3042d-c1df-7df2-86ff-ecfc5868dba1" [ 950.206974] env[65522]: _type = "Task" [ 950.206974] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.214838] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a3042d-c1df-7df2-86ff-ecfc5868dba1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.287836] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114386, 'name': Rename_Task, 'duration_secs': 0.183509} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.287836] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 950.288085] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab9ce9ec-3b75-4777-850c-ef5a7bedd1f5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.296108] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 950.296108] env[65522]: value = "task-5114389" [ 950.296108] env[65522]: _type = "Task" [ 950.296108] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.305521] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114389, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.354952] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114387, 'name': PowerOffVM_Task, 'duration_secs': 0.465444} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.355263] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 950.355480] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating instance 'a2a5fb2d-62e0-4809-a01f-f1df66dca58f' progress to 17 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 950.394845] env[65522]: DEBUG oslo_vmware.api [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114388, 'name': PowerOffVM_Task, 'duration_secs': 0.146957} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.395887] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.396583] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 950.396666] env[65522]: DEBUG nova.network.neutron [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 950.398866] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 950.398866] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 950.398866] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e102f7a-5003-41ac-afd7-ad8cc86eb13b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.426217] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 950.426465] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 950.426644] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Deleting the datastore file [datastore1] 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.426946] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc9e4ece-674e-490e-996c-947a2d17d4d0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.433926] env[65522]: DEBUG oslo_vmware.api [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Waiting for the task: (returnval){ [ 950.433926] env[65522]: value = "task-5114391" [ 950.433926] env[65522]: _type = "Task" [ 950.433926] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.443886] env[65522]: DEBUG oslo_vmware.api [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114391, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.553750] env[65522]: DEBUG nova.compute.manager [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 950.567400] env[65522]: DEBUG oslo_vmware.api [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Task: {'id': task-5114385, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.26527} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.567656] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.567841] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 950.568025] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 950.568199] env[65522]: INFO nova.compute.manager [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Took 2.16 seconds to destroy the instance on the hypervisor. [ 950.568464] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 950.568659] env[65522]: DEBUG nova.compute.manager [-] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 950.568886] env[65522]: DEBUG nova.network.neutron [-] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 950.569061] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 950.569607] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.569908] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.577073] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.051s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 950.580924] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.223s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 950.581143] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 950.582887] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.773s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 950.584445] env[65522]: INFO nova.compute.claims [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.587729] env[65522]: DEBUG oslo_concurrency.lockutils [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] Releasing lock "refresh_cache-8da2a500-6f0e-4eda-9a92-79510753b3b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 950.588008] env[65522]: DEBUG nova.compute.manager [req-8348a105-8205-4b14-b6bc-2d1f44fdf67e req-866de7ae-79ab-41c6-b70a-acc6e3295462 service nova] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Received event network-vif-deleted-7f6ad0cf-a0e9-4cb3-8a2d-3aa51e60b880 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 950.595482] env[65522]: DEBUG nova.virt.hardware [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 950.595706] env[65522]: DEBUG nova.virt.hardware [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 950.595896] env[65522]: DEBUG nova.virt.hardware [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 950.596902] env[65522]: DEBUG nova.virt.hardware [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 950.596902] env[65522]: DEBUG nova.virt.hardware [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 950.596902] env[65522]: DEBUG nova.virt.hardware [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 950.596902] env[65522]: DEBUG nova.virt.hardware [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.596902] env[65522]: DEBUG nova.virt.hardware [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 950.596902] env[65522]: DEBUG nova.virt.hardware [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 950.597240] env[65522]: DEBUG nova.virt.hardware [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 950.597240] env[65522]: DEBUG nova.virt.hardware [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 950.598481] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daee0f78-3ab3-4cef-87bc-cbc9493741be {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.608723] env[65522]: INFO nova.scheduler.client.report [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Deleted allocations for instance 42a08cfb-d865-4967-a086-370a2ca98b7d [ 950.610897] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f88068d-76cb-44ee-a282-8b7bfad0df2e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.615726] env[65522]: INFO nova.scheduler.client.report [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Deleted allocations for instance 5e227f6c-9f13-416f-8e6d-2f7d931619fd [ 950.623579] env[65522]: DEBUG nova.compute.manager [req-2eb71ccb-2db7-434d-80ac-435642827539 req-19e3533c-717c-4d89-9086-250b623c23be service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Received event network-vif-plugged-04e1485f-54a3-42d2-bd3f-19ea0fa7babe {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 950.623878] env[65522]: DEBUG oslo_concurrency.lockutils [req-2eb71ccb-2db7-434d-80ac-435642827539 req-19e3533c-717c-4d89-9086-250b623c23be service nova] Acquiring lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 950.624012] env[65522]: DEBUG oslo_concurrency.lockutils [req-2eb71ccb-2db7-434d-80ac-435642827539 req-19e3533c-717c-4d89-9086-250b623c23be service nova] Lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 950.624159] env[65522]: DEBUG oslo_concurrency.lockutils [req-2eb71ccb-2db7-434d-80ac-435642827539 req-19e3533c-717c-4d89-9086-250b623c23be service nova] Lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 950.624985] env[65522]: DEBUG nova.compute.manager [req-2eb71ccb-2db7-434d-80ac-435642827539 req-19e3533c-717c-4d89-9086-250b623c23be service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] No waiting events found dispatching network-vif-plugged-04e1485f-54a3-42d2-bd3f-19ea0fa7babe {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 950.624985] env[65522]: WARNING nova.compute.manager [req-2eb71ccb-2db7-434d-80ac-435642827539 req-19e3533c-717c-4d89-9086-250b623c23be service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Received unexpected event network-vif-plugged-04e1485f-54a3-42d2-bd3f-19ea0fa7babe for instance with vm_state active and task_state None. [ 950.665991] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 950.717446] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a3042d-c1df-7df2-86ff-ecfc5868dba1, 'name': SearchDatastore_Task, 'duration_secs': 0.018911} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.717765] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 950.718053] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 8da2a500-6f0e-4eda-9a92-79510753b3b8/8da2a500-6f0e-4eda-9a92-79510753b3b8.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 950.718376] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d4e1e72c-431f-4cd5-a9c0-ba948e253ba0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.726136] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 950.726136] env[65522]: value = "task-5114392" [ 950.726136] env[65522]: _type = "Task" [ 950.726136] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.735340] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114392, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.812013] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114389, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.862508] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 950.862820] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 950.863033] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 950.863259] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 950.863554] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 950.863795] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 950.864300] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.864633] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 950.864975] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 950.865209] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 950.865401] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 950.872324] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7eb5395b-64de-4a1b-9bbc-879c5cb71b69 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.893909] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 950.893909] env[65522]: value = "task-5114393" [ 950.893909] env[65522]: _type = "Task" [ 950.893909] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.900148] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.900661] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.913872] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114393, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.945242] env[65522]: DEBUG oslo_vmware.api [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Task: {'id': task-5114391, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.382218} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.945431] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.945570] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 950.945773] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 950.945964] env[65522]: INFO nova.compute.manager [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Took 1.09 seconds to destroy the instance on the hypervisor. [ 950.946445] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 950.946740] env[65522]: DEBUG nova.compute.manager [-] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 950.946925] env[65522]: DEBUG nova.network.neutron [-] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 950.947238] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 950.948158] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.948158] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.143742] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d78e4e18-e26f-47a8-acc5-a52e09da83c0 tempest-ServersTestBootFromVolume-414454397 tempest-ServersTestBootFromVolume-414454397-project-member] Lock "42a08cfb-d865-4967-a086-370a2ca98b7d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.909s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 951.145761] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5e15ab46-9a52-4271-83fd-447d2f557301 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "5e227f6c-9f13-416f-8e6d-2f7d931619fd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.918s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 951.242194] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114392, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.292017] env[65522]: WARNING nova.network.neutron [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] 1ea0ec37-ea40-4911-b6ad-54375253cc42 already exists in list: networks containing: ['1ea0ec37-ea40-4911-b6ad-54375253cc42']. ignoring it [ 951.292256] env[65522]: WARNING nova.network.neutron [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] 1ea0ec37-ea40-4911-b6ad-54375253cc42 already exists in list: networks containing: ['1ea0ec37-ea40-4911-b6ad-54375253cc42']. ignoring it [ 951.309016] env[65522]: DEBUG oslo_vmware.api [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114389, 'name': PowerOnVM_Task, 'duration_secs': 0.618292} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.309931] env[65522]: DEBUG nova.network.neutron [-] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 951.310199] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.311282] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 951.312563] env[65522]: INFO nova.compute.manager [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Took 10.00 seconds to spawn the instance on the hypervisor. [ 951.312563] env[65522]: DEBUG nova.compute.manager [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 951.312675] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c050343b-b8b5-4283-8f98-b20d8c2721a7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.389506] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.389874] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.399186] env[65522]: DEBUG nova.network.neutron [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Successfully updated port: bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 951.410397] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114393, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.457629] env[65522]: WARNING neutronclient.v2_0.client [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.458354] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.458682] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.566060] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.566468] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.629522] env[65522]: WARNING neutronclient.v2_0.client [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.630197] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.630609] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.668625] env[65522]: DEBUG nova.network.neutron [-] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 951.682226] env[65522]: DEBUG nova.compute.manager [req-b5a39248-801e-4ce8-b7b1-483111cbf2b1 req-4f0427c2-36d1-426d-91bd-c88c3294d871 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Received event network-vif-plugged-bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 951.682515] env[65522]: DEBUG oslo_concurrency.lockutils [req-b5a39248-801e-4ce8-b7b1-483111cbf2b1 req-4f0427c2-36d1-426d-91bd-c88c3294d871 service nova] Acquiring lock "79dd37b5-6b30-48ab-9f00-78214cbd132d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 951.682748] env[65522]: DEBUG oslo_concurrency.lockutils [req-b5a39248-801e-4ce8-b7b1-483111cbf2b1 req-4f0427c2-36d1-426d-91bd-c88c3294d871 service nova] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 951.682965] env[65522]: DEBUG oslo_concurrency.lockutils [req-b5a39248-801e-4ce8-b7b1-483111cbf2b1 req-4f0427c2-36d1-426d-91bd-c88c3294d871 service nova] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 951.683144] env[65522]: DEBUG nova.compute.manager [req-b5a39248-801e-4ce8-b7b1-483111cbf2b1 req-4f0427c2-36d1-426d-91bd-c88c3294d871 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] No waiting events found dispatching network-vif-plugged-bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 951.683308] env[65522]: WARNING nova.compute.manager [req-b5a39248-801e-4ce8-b7b1-483111cbf2b1 req-4f0427c2-36d1-426d-91bd-c88c3294d871 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Received unexpected event network-vif-plugged-bd432f78-055d-4cd1-b502-8e058700c03a for instance with vm_state building and task_state spawning. [ 951.739455] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114392, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550561} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.741367] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.741764] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.747936] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 8da2a500-6f0e-4eda-9a92-79510753b3b8/8da2a500-6f0e-4eda-9a92-79510753b3b8.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 951.748148] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 951.750795] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2d94b6c8-1d00-4953-a278-a3be6ba6d254 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.760182] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 951.760182] env[65522]: value = "task-5114394" [ 951.760182] env[65522]: _type = "Task" [ 951.760182] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.772419] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114394, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.814143] env[65522]: DEBUG nova.network.neutron [-] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 951.816914] env[65522]: WARNING neutronclient.v2_0.client [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.819135] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.819729] env[65522]: WARNING openstack [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.840027] env[65522]: INFO nova.compute.manager [-] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Took 0.89 seconds to deallocate network for instance. [ 951.844215] env[65522]: INFO nova.compute.manager [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Took 28.34 seconds to build instance. [ 951.904996] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.905194] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 951.905553] env[65522]: DEBUG nova.network.neutron [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 951.913466] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114393, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.957300] env[65522]: DEBUG nova.network.neutron [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updating instance_info_cache with network_info: [{"id": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "address": "fa:16:3e:8e:34:0c", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6221fb0-8c", "ovs_interfaceid": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d91694fd-ef8c-478e-8e0e-4e8ca56702a8", "address": "fa:16:3e:5d:61:15", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd91694fd-ef", "ovs_interfaceid": "d91694fd-ef8c-478e-8e0e-4e8ca56702a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "04e1485f-54a3-42d2-bd3f-19ea0fa7babe", "address": "fa:16:3e:fc:58:37", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04e1485f-54", "ovs_interfaceid": "04e1485f-54a3-42d2-bd3f-19ea0fa7babe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 952.022889] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb628b67-3f69-48d0-9edf-0ec1563a068b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.031456] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd90552f-4c6d-4717-9ce4-161c0a1d8c04 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.063946] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d57b0e2-3113-4398-8483-091b4a7ea2be {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.072504] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52fb0143-36c7-417a-9d7a-2f7bc439a179 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.086727] env[65522]: DEBUG nova.compute.provider_tree [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.171394] env[65522]: INFO nova.compute.manager [-] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Took 1.60 seconds to deallocate network for instance. [ 952.270779] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114394, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085579} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.271085] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 952.271927] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8aa291c-ba9b-4233-9e80-f06d3bfb1fe5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.296237] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 8da2a500-6f0e-4eda-9a92-79510753b3b8/8da2a500-6f0e-4eda-9a92-79510753b3b8.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.296755] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-127c3049-3cc9-46fe-b120-2f8a10ba81af {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.318677] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 952.318677] env[65522]: value = "task-5114395" [ 952.318677] env[65522]: _type = "Task" [ 952.318677] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.328434] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114395, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.347452] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 952.348106] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b66283b1-70db-417d-9a99-d136b94d1d36 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "84d4c689-b437-41a5-bb62-d144a30c6219" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.862s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 952.408142] env[65522]: WARNING openstack [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.408541] env[65522]: WARNING openstack [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.415649] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114393, 'name': ReconfigVM_Task, 'duration_secs': 1.46313} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.416192] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating instance 'a2a5fb2d-62e0-4809-a01f-f1df66dca58f' progress to 33 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 952.445800] env[65522]: DEBUG nova.network.neutron [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 952.460816] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 952.461529] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.461683] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 952.462885] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc386912-7c0b-43ee-a42a-3e4f5bd2dc31 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.467564] env[65522]: WARNING openstack [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.467923] env[65522]: WARNING openstack [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.491447] env[65522]: DEBUG nova.virt.hardware [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 952.491767] env[65522]: DEBUG nova.virt.hardware [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 952.491928] env[65522]: DEBUG nova.virt.hardware [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 952.492123] env[65522]: DEBUG nova.virt.hardware [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 952.492265] env[65522]: DEBUG nova.virt.hardware [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 952.492408] env[65522]: DEBUG nova.virt.hardware [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 952.492606] env[65522]: DEBUG nova.virt.hardware [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 952.492759] env[65522]: DEBUG nova.virt.hardware [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 952.492919] env[65522]: DEBUG nova.virt.hardware [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 952.493104] env[65522]: DEBUG nova.virt.hardware [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 952.493280] env[65522]: DEBUG nova.virt.hardware [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 952.499544] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Reconfiguring VM to attach interface {{(pid=65522) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 952.500501] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d389cb97-2028-44af-b7cd-5172e3287a4e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.524097] env[65522]: DEBUG oslo_vmware.api [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 952.524097] env[65522]: value = "task-5114396" [ 952.524097] env[65522]: _type = "Task" [ 952.524097] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.533884] env[65522]: DEBUG oslo_vmware.api [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114396, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.561970] env[65522]: WARNING neutronclient.v2_0.client [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 952.562674] env[65522]: WARNING openstack [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.563045] env[65522]: WARNING openstack [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.590701] env[65522]: DEBUG nova.scheduler.client.report [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 952.664376] env[65522]: DEBUG nova.network.neutron [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Updating instance_info_cache with network_info: [{"id": "bd432f78-055d-4cd1-b502-8e058700c03a", "address": "fa:16:3e:38:8a:d1", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd432f78-05", "ovs_interfaceid": "bd432f78-055d-4cd1-b502-8e058700c03a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 952.680772] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 952.837020] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114395, 'name': ReconfigVM_Task, 'duration_secs': 0.428774} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.837020] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 8da2a500-6f0e-4eda-9a92-79510753b3b8/8da2a500-6f0e-4eda-9a92-79510753b3b8.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 952.837020] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bfedb0cd-14e5-496b-b6ed-f13dc9ef601b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.843678] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 952.843678] env[65522]: value = "task-5114397" [ 952.843678] env[65522]: _type = "Task" [ 952.843678] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.853399] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114397, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.910541] env[65522]: DEBUG nova.compute.manager [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Received event network-changed-04e1485f-54a3-42d2-bd3f-19ea0fa7babe {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 952.910744] env[65522]: DEBUG nova.compute.manager [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Refreshing instance network info cache due to event network-changed-04e1485f-54a3-42d2-bd3f-19ea0fa7babe. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 952.910960] env[65522]: DEBUG oslo_concurrency.lockutils [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] Acquiring lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.911135] env[65522]: DEBUG oslo_concurrency.lockutils [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] Acquired lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 952.911301] env[65522]: DEBUG nova.network.neutron [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Refreshing network info cache for port 04e1485f-54a3-42d2-bd3f-19ea0fa7babe {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 952.923669] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 952.923669] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 952.923821] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 952.923956] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 952.924118] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 952.924258] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 952.924447] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 952.924590] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 952.926910] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 952.926910] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 952.926910] env[65522]: DEBUG nova.virt.hardware [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 952.930560] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Reconfiguring VM instance instance-00000044 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 952.932985] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa53ee07-97c2-46aa-8dcf-7588728a4804 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.952282] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 952.952282] env[65522]: value = "task-5114398" [ 952.952282] env[65522]: _type = "Task" [ 952.952282] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.956169] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b402f050-ae56-4fb3-9507-ba8a09cbd57e tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "84d4c689-b437-41a5-bb62-d144a30c6219" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 952.956414] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b402f050-ae56-4fb3-9507-ba8a09cbd57e tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "84d4c689-b437-41a5-bb62-d144a30c6219" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 952.956591] env[65522]: DEBUG nova.compute.manager [None req-b402f050-ae56-4fb3-9507-ba8a09cbd57e tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 952.957426] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bb5a67-7592-4100-98cf-79f4f2065d86 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.967806] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114398, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.969663] env[65522]: DEBUG nova.compute.manager [None req-b402f050-ae56-4fb3-9507-ba8a09cbd57e tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65522) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 952.970273] env[65522]: DEBUG nova.objects.instance [None req-b402f050-ae56-4fb3-9507-ba8a09cbd57e tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lazy-loading 'flavor' on Instance uuid 84d4c689-b437-41a5-bb62-d144a30c6219 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.037092] env[65522]: DEBUG oslo_vmware.api [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114396, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.096344] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 953.097031] env[65522]: DEBUG nova.compute.manager [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 953.101388] env[65522]: DEBUG oslo_concurrency.lockutils [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.013s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 953.101580] env[65522]: DEBUG nova.objects.instance [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lazy-loading 'resources' on Instance uuid 0b48b747-149c-4163-bdd0-7d0b07ddacfd {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.170682] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 953.170682] env[65522]: DEBUG nova.compute.manager [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Instance network_info: |[{"id": "bd432f78-055d-4cd1-b502-8e058700c03a", "address": "fa:16:3e:38:8a:d1", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd432f78-05", "ovs_interfaceid": "bd432f78-055d-4cd1-b502-8e058700c03a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 953.170682] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:8a:d1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4c5eb94-841c-4713-985a-8fc4117fbaf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd432f78-055d-4cd1-b502-8e058700c03a', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 953.179924] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Creating folder: Project (1546bbdca07c443d84abea1155cfde37). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 953.180526] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.180835] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 953.181102] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.181334] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 953.181579] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 953.183904] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3cda489b-280f-4604-9913-8ffb79d0e90f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.187073] env[65522]: INFO nova.compute.manager [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Terminating instance [ 953.202091] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Created folder: Project (1546bbdca07c443d84abea1155cfde37) in parent group-v994660. [ 953.202091] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Creating folder: Instances. Parent ref: group-v994861. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 953.202091] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c9d2fc8c-ea35-4b20-92cf-2556ef843660 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.214256] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Created folder: Instances in parent group-v994861. [ 953.214390] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 953.214692] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 953.215039] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4e1aeb5-9ecc-4024-b0af-fa5598254291 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.237952] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 953.237952] env[65522]: value = "task-5114401" [ 953.237952] env[65522]: _type = "Task" [ 953.237952] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.248504] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114401, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.355634] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114397, 'name': Rename_Task, 'duration_secs': 0.308413} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.355944] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 953.356224] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-834d3dd6-081f-4a22-952a-62339f5e7892 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.364764] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 953.364764] env[65522]: value = "task-5114402" [ 953.364764] env[65522]: _type = "Task" [ 953.364764] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.374089] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114402, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.413943] env[65522]: WARNING neutronclient.v2_0.client [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.414716] env[65522]: WARNING openstack [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.415164] env[65522]: WARNING openstack [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.464591] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114398, 'name': ReconfigVM_Task, 'duration_secs': 0.216544} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.464894] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Reconfigured VM instance instance-00000044 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 953.467021] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6249986-6d8f-4c14-9e0a-0ce7ac58a51b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.498555] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] a2a5fb2d-62e0-4809-a01f-f1df66dca58f/a2a5fb2d-62e0-4809-a01f-f1df66dca58f.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 953.500408] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d239c9cf-3707-40e2-a763-255b5e68be81 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.531642] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 953.531642] env[65522]: value = "task-5114403" [ 953.531642] env[65522]: _type = "Task" [ 953.531642] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.545585] env[65522]: DEBUG oslo_vmware.api [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114396, 'name': ReconfigVM_Task, 'duration_secs': 0.684783} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.547330] env[65522]: WARNING openstack [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.547840] env[65522]: WARNING openstack [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.557118] env[65522]: WARNING neutronclient.v2_0.client [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.557381] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 953.557590] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Reconfigured VM to attach interface {{(pid=65522) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 953.564319] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114403, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.606546] env[65522]: DEBUG nova.compute.utils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 953.612282] env[65522]: DEBUG nova.compute.manager [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 953.612526] env[65522]: DEBUG nova.network.neutron [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 953.612884] env[65522]: WARNING neutronclient.v2_0.client [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.613236] env[65522]: WARNING neutronclient.v2_0.client [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.613991] env[65522]: WARNING openstack [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.614344] env[65522]: WARNING openstack [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.658504] env[65522]: WARNING neutronclient.v2_0.client [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.659236] env[65522]: WARNING openstack [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 953.659606] env[65522]: WARNING openstack [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 953.690037] env[65522]: DEBUG nova.policy [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a97b2f232f7844c993dccdc386ce3b2c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9614f18e01594bf083e619c11760fe36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 953.692198] env[65522]: DEBUG nova.compute.manager [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 953.692734] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 953.693840] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a1ae8c-0653-4f7d-8f61-8069c580c0c7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.704745] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.705192] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41fc7c37-c78d-413d-8b41-c1bd383c36f0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.713546] env[65522]: DEBUG oslo_vmware.api [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 953.713546] env[65522]: value = "task-5114404" [ 953.713546] env[65522]: _type = "Task" [ 953.713546] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.726665] env[65522]: DEBUG oslo_vmware.api [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114404, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.753112] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114401, 'name': CreateVM_Task, 'duration_secs': 0.411156} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.753339] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 953.753893] env[65522]: WARNING neutronclient.v2_0.client [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 953.754298] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.754462] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 953.754823] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 953.755101] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdb1b4bd-47d8-47e8-b6f3-1bca71026a78 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.765282] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 953.765282] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a83bfc-17bd-ca0a-be09-17a90b546e0d" [ 953.765282] env[65522]: _type = "Task" [ 953.765282] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.779149] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a83bfc-17bd-ca0a-be09-17a90b546e0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.882261] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114402, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.977026] env[65522]: DEBUG nova.network.neutron [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updated VIF entry in instance network info cache for port 04e1485f-54a3-42d2-bd3f-19ea0fa7babe. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 953.977695] env[65522]: DEBUG nova.network.neutron [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updating instance_info_cache with network_info: [{"id": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "address": "fa:16:3e:8e:34:0c", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6221fb0-8c", "ovs_interfaceid": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d91694fd-ef8c-478e-8e0e-4e8ca56702a8", "address": "fa:16:3e:5d:61:15", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd91694fd-ef", "ovs_interfaceid": "d91694fd-ef8c-478e-8e0e-4e8ca56702a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "04e1485f-54a3-42d2-bd3f-19ea0fa7babe", "address": "fa:16:3e:fc:58:37", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04e1485f-54", "ovs_interfaceid": "04e1485f-54a3-42d2-bd3f-19ea0fa7babe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 953.979457] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b402f050-ae56-4fb3-9507-ba8a09cbd57e tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.979751] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2e7ffc3-174b-4fb7-9795-3b8735a734a3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.990037] env[65522]: DEBUG oslo_vmware.api [None req-b402f050-ae56-4fb3-9507-ba8a09cbd57e tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 953.990037] env[65522]: value = "task-5114405" [ 953.990037] env[65522]: _type = "Task" [ 953.990037] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.003221] env[65522]: DEBUG oslo_vmware.api [None req-b402f050-ae56-4fb3-9507-ba8a09cbd57e tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114405, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.050742] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114403, 'name': ReconfigVM_Task, 'duration_secs': 0.43832} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.051634] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Reconfigured VM instance instance-00000044 to attach disk [datastore1] a2a5fb2d-62e0-4809-a01f-f1df66dca58f/a2a5fb2d-62e0-4809-a01f-f1df66dca58f.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 954.051634] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating instance 'a2a5fb2d-62e0-4809-a01f-f1df66dca58f' progress to 50 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 954.066733] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a823daaa-b72a-40e3-9c8b-3d94a3877158 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-9df1a51a-2811-4486-a4c6-58d618f2ae7d-04e1485f-54a3-42d2-bd3f-19ea0fa7babe" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.070s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.075163] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e665ac6e-3fe1-4a7d-b264-b7293756db73 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.085955] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b0cd2d-46f0-434f-b133-fa2d5f683ead {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.127811] env[65522]: DEBUG nova.compute.manager [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 954.132085] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf41de24-8a46-4117-a088-74a5c1e07145 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.138567] env[65522]: DEBUG nova.compute.manager [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Received event network-changed-bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 954.138864] env[65522]: DEBUG nova.compute.manager [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Refreshing instance network info cache due to event network-changed-bd432f78-055d-4cd1-b502-8e058700c03a. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 954.139430] env[65522]: DEBUG oslo_concurrency.lockutils [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] Acquiring lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.139430] env[65522]: DEBUG oslo_concurrency.lockutils [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] Acquired lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 954.139516] env[65522]: DEBUG nova.network.neutron [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Refreshing network info cache for port bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 954.148392] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e452ef-a9f3-4a3d-9f47-e87d9db50362 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.168558] env[65522]: DEBUG nova.compute.provider_tree [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.192859] env[65522]: DEBUG nova.network.neutron [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Successfully created port: 626b09b1-0266-4a08-a37f-135470571842 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 954.228598] env[65522]: DEBUG oslo_vmware.api [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114404, 'name': PowerOffVM_Task, 'duration_secs': 0.3243} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.229074] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.229074] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 954.229501] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d0f14a5a-a6b5-4054-92de-6244adcf47d4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.276833] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a83bfc-17bd-ca0a-be09-17a90b546e0d, 'name': SearchDatastore_Task, 'duration_secs': 0.013523} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.277524] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 954.277524] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 954.277755] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.279499] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 954.279499] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 954.279499] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-274f511c-4af2-47d9-b239-4125101598fa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.289392] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 954.289568] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 954.290694] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49595e5d-c97e-4422-b50c-0c52b5c79acf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.297741] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 954.297741] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ddda68-41c2-502b-ba8e-3d16525a8592" [ 954.297741] env[65522]: _type = "Task" [ 954.297741] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.311179] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ddda68-41c2-502b-ba8e-3d16525a8592, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.319171] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 954.319405] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 954.319581] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Deleting the datastore file [datastore1] d5e8dd05-dc3c-4831-b4b0-ac100360f3e7 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.320161] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2741a572-1449-4d4c-9dc6-279714d3bf07 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.328662] env[65522]: DEBUG oslo_vmware.api [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for the task: (returnval){ [ 954.328662] env[65522]: value = "task-5114407" [ 954.328662] env[65522]: _type = "Task" [ 954.328662] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.340430] env[65522]: DEBUG oslo_vmware.api [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114407, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.376845] env[65522]: DEBUG oslo_vmware.api [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114402, 'name': PowerOnVM_Task, 'duration_secs': 0.617532} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.377982] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 954.377982] env[65522]: INFO nova.compute.manager [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Took 9.01 seconds to spawn the instance on the hypervisor. [ 954.377982] env[65522]: DEBUG nova.compute.manager [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 954.378413] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a3206a-cf06-468c-b2ca-4731f2366c49 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.482994] env[65522]: DEBUG oslo_concurrency.lockutils [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] Releasing lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 954.487350] env[65522]: DEBUG nova.compute.manager [req-829ae844-12e9-4bac-b32b-5cfdfb8c7ea1 req-65f609ac-e419-4aef-8a61-c1ce5e96c85d service nova] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Received event network-vif-deleted-c7aacbb2-e620-48af-838a-6103f858f4c8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 954.501701] env[65522]: DEBUG oslo_vmware.api [None req-b402f050-ae56-4fb3-9507-ba8a09cbd57e tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114405, 'name': PowerOffVM_Task, 'duration_secs': 0.288834} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.501999] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b402f050-ae56-4fb3-9507-ba8a09cbd57e tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.502197] env[65522]: DEBUG nova.compute.manager [None req-b402f050-ae56-4fb3-9507-ba8a09cbd57e tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 954.503070] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28760b7b-a5ec-4cff-bfdc-47caf3c91f33 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.564332] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c047a17-3ea0-4215-9c25-86fcdf4997a1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.587829] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88bf7021-eb2b-4062-8a3c-51f89ad824d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.608500] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating instance 'a2a5fb2d-62e0-4809-a01f-f1df66dca58f' progress to 67 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 954.644031] env[65522]: WARNING neutronclient.v2_0.client [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 954.644582] env[65522]: WARNING openstack [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 954.646026] env[65522]: WARNING openstack [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 954.672896] env[65522]: DEBUG nova.scheduler.client.report [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 954.790926] env[65522]: WARNING openstack [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 954.791425] env[65522]: WARNING openstack [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 954.810978] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ddda68-41c2-502b-ba8e-3d16525a8592, 'name': SearchDatastore_Task, 'duration_secs': 0.025168} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.812129] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c2b12fd-e087-49c5-afca-9f021b60bf3f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.819662] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 954.819662] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f06490-23fe-fe12-6691-5f6ea49035db" [ 954.819662] env[65522]: _type = "Task" [ 954.819662] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.832107] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f06490-23fe-fe12-6691-5f6ea49035db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.842653] env[65522]: DEBUG oslo_vmware.api [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114407, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.888360] env[65522]: WARNING neutronclient.v2_0.client [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 954.889273] env[65522]: WARNING openstack [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 954.889807] env[65522]: WARNING openstack [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 954.913887] env[65522]: INFO nova.compute.manager [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Took 29.49 seconds to build instance. [ 954.992690] env[65522]: DEBUG nova.network.neutron [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Updated VIF entry in instance network info cache for port bd432f78-055d-4cd1-b502-8e058700c03a. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 954.993071] env[65522]: DEBUG nova.network.neutron [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Updating instance_info_cache with network_info: [{"id": "bd432f78-055d-4cd1-b502-8e058700c03a", "address": "fa:16:3e:38:8a:d1", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd432f78-05", "ovs_interfaceid": "bd432f78-055d-4cd1-b502-8e058700c03a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 955.017035] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b402f050-ae56-4fb3-9507-ba8a09cbd57e tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "84d4c689-b437-41a5-bb62-d144a30c6219" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.060s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 955.114971] env[65522]: WARNING neutronclient.v2_0.client [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 955.143070] env[65522]: DEBUG nova.compute.manager [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 955.171612] env[65522]: DEBUG nova.virt.hardware [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 955.171833] env[65522]: DEBUG nova.virt.hardware [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 955.171991] env[65522]: DEBUG nova.virt.hardware [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 955.172184] env[65522]: DEBUG nova.virt.hardware [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 955.172328] env[65522]: DEBUG nova.virt.hardware [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 955.172473] env[65522]: DEBUG nova.virt.hardware [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 955.172718] env[65522]: DEBUG nova.virt.hardware [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 955.172869] env[65522]: DEBUG nova.virt.hardware [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 955.172979] env[65522]: DEBUG nova.virt.hardware [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 955.173154] env[65522]: DEBUG nova.virt.hardware [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 955.173325] env[65522]: DEBUG nova.virt.hardware [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 955.174271] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86da06d8-313e-4ecb-9fc8-54dd657ee5e2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.177771] env[65522]: DEBUG nova.network.neutron [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Port f8145f30-4e64-4a22-89a7-ffa393891ffb binding to destination host cpu-1 is already ACTIVE {{(pid=65522) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 955.179652] env[65522]: DEBUG oslo_concurrency.lockutils [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.078s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 955.183120] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.866s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 955.183120] env[65522]: DEBUG nova.objects.instance [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Lazy-loading 'resources' on Instance uuid 2a12a0a3-913e-4ade-85ef-cca7ba508ac9 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.188494] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9659d2d3-2e56-4eed-a20f-cbecbde0677a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.213449] env[65522]: INFO nova.scheduler.client.report [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleted allocations for instance 0b48b747-149c-4163-bdd0-7d0b07ddacfd [ 955.331950] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f06490-23fe-fe12-6691-5f6ea49035db, 'name': SearchDatastore_Task, 'duration_secs': 0.023103} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.332342] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 955.332597] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 79dd37b5-6b30-48ab-9f00-78214cbd132d/79dd37b5-6b30-48ab-9f00-78214cbd132d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 955.336116] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-69b5c085-6df8-4d08-88b4-10b66bd89cd8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.346067] env[65522]: DEBUG oslo_vmware.api [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Task: {'id': task-5114407, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.655781} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.346514] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.346955] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 955.346955] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 955.347136] env[65522]: INFO nova.compute.manager [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Took 1.65 seconds to destroy the instance on the hypervisor. [ 955.347339] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 955.347603] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 955.347603] env[65522]: value = "task-5114408" [ 955.347603] env[65522]: _type = "Task" [ 955.347603] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.347799] env[65522]: DEBUG nova.compute.manager [-] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 955.347895] env[65522]: DEBUG nova.network.neutron [-] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 955.348158] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 955.348681] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.348974] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.365841] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114408, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.417319] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a9a809a-741f-4014-9a7f-65e1352c9efc tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "8da2a500-6f0e-4eda-9a92-79510753b3b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.006s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 955.423676] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 955.496108] env[65522]: DEBUG oslo_concurrency.lockutils [req-8a8d2b3b-f471-41b6-ac0a-24fa98766666 req-e76c692c-a6d4-4d97-aad1-713d4ea6757c service nova] Releasing lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 955.724789] env[65522]: DEBUG oslo_concurrency.lockutils [None req-22382972-f371-45da-b834-d0ed0cda2d82 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "0b48b747-149c-4163-bdd0-7d0b07ddacfd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.672s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 955.870624] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114408, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.872672] env[65522]: DEBUG nova.network.neutron [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Successfully updated port: 626b09b1-0266-4a08-a37f-135470571842 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 955.874847] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 955.874847] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 955.955683] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5689b54-4bbf-4432-a0af-8e20d882c5cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.968629] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-90ab22f1-db4d-4e73-8e2b-e8794a30734f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Suspending the VM {{(pid=65522) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 955.969344] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-28699e7f-f8cb-4d01-b113-a1ca74e11d3e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.981236] env[65522]: DEBUG oslo_vmware.api [None req-90ab22f1-db4d-4e73-8e2b-e8794a30734f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 955.981236] env[65522]: value = "task-5114409" [ 955.981236] env[65522]: _type = "Task" [ 955.981236] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.992289] env[65522]: DEBUG oslo_vmware.api [None req-90ab22f1-db4d-4e73-8e2b-e8794a30734f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114409, 'name': SuspendVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.038799] env[65522]: DEBUG nova.compute.manager [req-5760d1f0-87ed-4b0c-9cd4-0f25bea11e9d req-c8d65d63-b5f7-4066-bcac-478d04b464c3 service nova] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Received event network-vif-deleted-fb1cd05a-757b-469c-9749-fd1ca4fcb168 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 956.038991] env[65522]: INFO nova.compute.manager [req-5760d1f0-87ed-4b0c-9cd4-0f25bea11e9d req-c8d65d63-b5f7-4066-bcac-478d04b464c3 service nova] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Neutron deleted interface fb1cd05a-757b-469c-9749-fd1ca4fcb168; detaching it from the instance and deleting it from the info cache [ 956.039202] env[65522]: DEBUG nova.network.neutron [req-5760d1f0-87ed-4b0c-9cd4-0f25bea11e9d req-c8d65d63-b5f7-4066-bcac-478d04b464c3 service nova] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 956.096542] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233212d7-94c9-4ead-9e0f-647340e02802 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.109505] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba8485a-5e5e-4a2a-8a24-a7b45aea4772 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.147690] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa7fdd3b-e4d1-43de-bda8-6d6e529cbb3e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.151172] env[65522]: DEBUG nova.network.neutron [-] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 956.161601] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6867c4c-c4a9-44e0-9b14-c5fb37b186a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.180071] env[65522]: DEBUG nova.compute.provider_tree [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.210321] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 956.210963] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.210963] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 956.370711] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114408, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.761787} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.371617] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 79dd37b5-6b30-48ab-9f00-78214cbd132d/79dd37b5-6b30-48ab-9f00-78214cbd132d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 956.371847] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.372244] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d7a4c65-2bae-4968-bc6a-51938dc1848d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.380328] env[65522]: INFO nova.compute.manager [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Detaching volume 33425e73-8f97-4928-afed-0733d11034fb [ 956.385785] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "refresh_cache-1b2779f5-c6e8-4226-a819-0560d63bd7df" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.385785] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "refresh_cache-1b2779f5-c6e8-4226-a819-0560d63bd7df" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 956.385785] env[65522]: DEBUG nova.network.neutron [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 956.388338] env[65522]: DEBUG nova.compute.manager [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Received event network-vif-plugged-626b09b1-0266-4a08-a37f-135470571842 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 956.388684] env[65522]: DEBUG oslo_concurrency.lockutils [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] Acquiring lock "1b2779f5-c6e8-4226-a819-0560d63bd7df-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 956.388885] env[65522]: DEBUG oslo_concurrency.lockutils [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] Lock "1b2779f5-c6e8-4226-a819-0560d63bd7df-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.389214] env[65522]: DEBUG oslo_concurrency.lockutils [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] Lock "1b2779f5-c6e8-4226-a819-0560d63bd7df-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 956.389214] env[65522]: DEBUG nova.compute.manager [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] No waiting events found dispatching network-vif-plugged-626b09b1-0266-4a08-a37f-135470571842 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 956.389372] env[65522]: WARNING nova.compute.manager [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Received unexpected event network-vif-plugged-626b09b1-0266-4a08-a37f-135470571842 for instance with vm_state building and task_state spawning. [ 956.389719] env[65522]: DEBUG nova.compute.manager [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Received event network-changed-626b09b1-0266-4a08-a37f-135470571842 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 956.389719] env[65522]: DEBUG nova.compute.manager [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Refreshing instance network info cache due to event network-changed-626b09b1-0266-4a08-a37f-135470571842. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 956.390218] env[65522]: DEBUG oslo_concurrency.lockutils [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] Acquiring lock "refresh_cache-1b2779f5-c6e8-4226-a819-0560d63bd7df" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.391032] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 956.391032] env[65522]: value = "task-5114410" [ 956.391032] env[65522]: _type = "Task" [ 956.391032] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.406553] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114410, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.429874] env[65522]: INFO nova.virt.block_device [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Attempting to driver detach volume 33425e73-8f97-4928-afed-0733d11034fb from mountpoint /dev/sdb [ 956.430809] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 956.430809] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994827', 'volume_id': '33425e73-8f97-4928-afed-0733d11034fb', 'name': 'volume-33425e73-8f97-4928-afed-0733d11034fb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1eebb320-30e2-4d8b-b1fd-19e31fb35b77', 'attached_at': '', 'detached_at': '', 'volume_id': '33425e73-8f97-4928-afed-0733d11034fb', 'serial': '33425e73-8f97-4928-afed-0733d11034fb'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 956.431319] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b44ec5-a2e6-49e0-8a41-d178fe96222f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.456464] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c2725f-3707-4b4d-9591-1b41745baeb2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.465617] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e37d7198-ceeb-4c4b-a702-c06996f190f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.490880] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd209038-21b3-4f8a-9453-c3541d0ae313 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.512220] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] The volume has not been displaced from its original location: [datastore2] volume-33425e73-8f97-4928-afed-0733d11034fb/volume-33425e73-8f97-4928-afed-0733d11034fb.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 956.517737] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Reconfiguring VM instance instance-0000002e to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 956.522463] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39760bf9-72ac-4e62-9f89-113c936c638c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.538901] env[65522]: DEBUG oslo_vmware.api [None req-90ab22f1-db4d-4e73-8e2b-e8794a30734f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114409, 'name': SuspendVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.549593] env[65522]: DEBUG oslo_vmware.api [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 956.549593] env[65522]: value = "task-5114411" [ 956.549593] env[65522]: _type = "Task" [ 956.549593] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.549848] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4ff9550b-6a02-4546-a29e-fb9b1c44a36c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.563581] env[65522]: DEBUG oslo_vmware.api [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114411, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.567713] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931459a5-fd71-4c0f-a683-0cc8c872d2c3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.609486] env[65522]: DEBUG nova.compute.manager [req-5760d1f0-87ed-4b0c-9cd4-0f25bea11e9d req-c8d65d63-b5f7-4066-bcac-478d04b464c3 service nova] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Detach interface failed, port_id=fb1cd05a-757b-469c-9749-fd1ca4fcb168, reason: Instance d5e8dd05-dc3c-4831-b4b0-ac100360f3e7 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 956.654818] env[65522]: INFO nova.compute.manager [-] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Took 1.31 seconds to deallocate network for instance. [ 956.683318] env[65522]: DEBUG nova.scheduler.client.report [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 956.806082] env[65522]: INFO nova.compute.manager [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Rebuilding instance [ 956.853558] env[65522]: DEBUG nova.compute.manager [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 956.854537] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe411398-6d07-4c1c-b793-67cfc54368a0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.860676] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "interface-9df1a51a-2811-4486-a4c6-58d618f2ae7d-d91694fd-ef8c-478e-8e0e-4e8ca56702a8" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 956.860933] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-9df1a51a-2811-4486-a4c6-58d618f2ae7d-d91694fd-ef8c-478e-8e0e-4e8ca56702a8" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.895166] env[65522]: WARNING openstack [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 956.895560] env[65522]: WARNING openstack [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 956.913448] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114410, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088563} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.913962] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 956.914637] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea2ab6d-eaaf-494d-accc-4ba9371a9282 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.940648] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 79dd37b5-6b30-48ab-9f00-78214cbd132d/79dd37b5-6b30-48ab-9f00-78214cbd132d.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.941776] env[65522]: DEBUG nova.network.neutron [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 956.943917] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42dca784-0b19-420d-80e1-660bdf66ec8a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.964898] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 956.964898] env[65522]: value = "task-5114412" [ 956.964898] env[65522]: _type = "Task" [ 956.964898] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.973839] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114412, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.976250] env[65522]: WARNING openstack [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 956.976631] env[65522]: WARNING openstack [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 956.996812] env[65522]: DEBUG oslo_vmware.api [None req-90ab22f1-db4d-4e73-8e2b-e8794a30734f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114409, 'name': SuspendVM_Task, 'duration_secs': 0.830236} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.996812] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-90ab22f1-db4d-4e73-8e2b-e8794a30734f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Suspended the VM {{(pid=65522) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 956.996812] env[65522]: DEBUG nova.compute.manager [None req-90ab22f1-db4d-4e73-8e2b-e8794a30734f tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 956.997311] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c633a0a-f295-4b4a-b8b9-a57dfab8190c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.040350] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "24f520ec-6a56-4f17-9ae4-2c856bf99582" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.041030] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "24f520ec-6a56-4f17-9ae4-2c856bf99582" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 957.041125] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "24f520ec-6a56-4f17-9ae4-2c856bf99582-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.041389] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "24f520ec-6a56-4f17-9ae4-2c856bf99582-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 957.041653] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "24f520ec-6a56-4f17-9ae4-2c856bf99582-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 957.044267] env[65522]: INFO nova.compute.manager [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Terminating instance [ 957.046969] env[65522]: WARNING neutronclient.v2_0.client [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.047553] env[65522]: WARNING openstack [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.047897] env[65522]: WARNING openstack [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.067876] env[65522]: DEBUG oslo_vmware.api [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114411, 'name': ReconfigVM_Task, 'duration_secs': 0.295191} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.068094] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Reconfigured VM instance instance-0000002e to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 957.073226] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-085f0de1-b520-4300-a188-175e229ba26e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.090425] env[65522]: DEBUG oslo_vmware.api [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 957.090425] env[65522]: value = "task-5114413" [ 957.090425] env[65522]: _type = "Task" [ 957.090425] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.104958] env[65522]: DEBUG oslo_vmware.api [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114413, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.138106] env[65522]: DEBUG nova.network.neutron [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Updating instance_info_cache with network_info: [{"id": "626b09b1-0266-4a08-a37f-135470571842", "address": "fa:16:3e:cb:66:49", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap626b09b1-02", "ovs_interfaceid": "626b09b1-0266-4a08-a37f-135470571842", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 957.165090] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.189122] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.006s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 957.192149] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.532s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 957.193838] env[65522]: INFO nova.compute.claims [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 957.216999] env[65522]: INFO nova.scheduler.client.report [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Deleted allocations for instance 2a12a0a3-913e-4ade-85ef-cca7ba508ac9 [ 957.219752] env[65522]: WARNING neutronclient.v2_0.client [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.263764] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.263945] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 957.264130] env[65522]: DEBUG nova.network.neutron [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 957.367245] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.367445] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 957.370268] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ed6a719-041b-4d13-aabc-e61da0dbffb1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.390054] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472b21a9-7ea1-42ef-80bc-78c8bf893bb2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.418065] env[65522]: WARNING neutronclient.v2_0.client [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.424852] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Reconfiguring VM to detach interface {{(pid=65522) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 957.425237] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cb40a82-4c89-4c79-85a3-6b61c75bd3bb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.445480] env[65522]: DEBUG oslo_vmware.api [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 957.445480] env[65522]: value = "task-5114414" [ 957.445480] env[65522]: _type = "Task" [ 957.445480] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.454985] env[65522]: DEBUG oslo_vmware.api [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114414, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.476252] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114412, 'name': ReconfigVM_Task, 'duration_secs': 0.389709} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.476540] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 79dd37b5-6b30-48ab-9f00-78214cbd132d/79dd37b5-6b30-48ab-9f00-78214cbd132d.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.478035] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0814352-ed8c-46f3-984c-22e2c795cd13 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.485376] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 957.485376] env[65522]: value = "task-5114415" [ 957.485376] env[65522]: _type = "Task" [ 957.485376] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.495783] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114415, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.561817] env[65522]: DEBUG nova.compute.manager [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 957.561990] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 957.562935] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26f8cc5-d69f-457a-a0fd-d14f15b98472 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.572170] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 957.572445] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a58801ca-9c6a-48ae-be86-f1059e92030c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.581171] env[65522]: DEBUG oslo_vmware.api [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 957.581171] env[65522]: value = "task-5114416" [ 957.581171] env[65522]: _type = "Task" [ 957.581171] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.590802] env[65522]: DEBUG oslo_vmware.api [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114416, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.600469] env[65522]: DEBUG oslo_vmware.api [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114413, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.641602] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "refresh_cache-1b2779f5-c6e8-4226-a819-0560d63bd7df" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 957.642010] env[65522]: DEBUG nova.compute.manager [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Instance network_info: |[{"id": "626b09b1-0266-4a08-a37f-135470571842", "address": "fa:16:3e:cb:66:49", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap626b09b1-02", "ovs_interfaceid": "626b09b1-0266-4a08-a37f-135470571842", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 957.642378] env[65522]: DEBUG oslo_concurrency.lockutils [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] Acquired lock "refresh_cache-1b2779f5-c6e8-4226-a819-0560d63bd7df" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 957.642553] env[65522]: DEBUG nova.network.neutron [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Refreshing network info cache for port 626b09b1-0266-4a08-a37f-135470571842 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 957.643824] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:66:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10b81051-1eb1-406b-888c-4548c470c77e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '626b09b1-0266-4a08-a37f-135470571842', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 957.652455] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 957.653924] env[65522]: WARNING neutronclient.v2_0.client [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.654845] env[65522]: WARNING openstack [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.654928] env[65522]: WARNING openstack [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.662447] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 957.663663] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7088b2cd-7ca3-454f-ba38-3407fb21b002 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.686024] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 957.686024] env[65522]: value = "task-5114417" [ 957.686024] env[65522]: _type = "Task" [ 957.686024] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.695068] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114417, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.728497] env[65522]: DEBUG oslo_concurrency.lockutils [None req-10377fd7-897e-4f06-b9ab-030a502cdc50 tempest-ServerPasswordTestJSON-784794184 tempest-ServerPasswordTestJSON-784794184-project-member] Lock "2a12a0a3-913e-4ade-85ef-cca7ba508ac9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.203s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 957.767399] env[65522]: WARNING neutronclient.v2_0.client [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.767995] env[65522]: WARNING openstack [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.768361] env[65522]: WARNING openstack [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.874829] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 957.875203] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-335e8c49-e67b-4f18-8d5c-3ebcc312213c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.887561] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 957.887561] env[65522]: value = "task-5114418" [ 957.887561] env[65522]: _type = "Task" [ 957.887561] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.900628] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 957.900864] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 957.901689] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb82197-a7a4-449e-837b-9c9ee9d6a914 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.905896] env[65522]: WARNING openstack [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.906274] env[65522]: WARNING openstack [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.920037] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 957.920290] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8d6fd21a-ff9d-477d-a412-b790bd9fc90f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.958598] env[65522]: DEBUG oslo_vmware.api [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114414, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.992579] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 957.992766] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 957.992952] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleting the datastore file [datastore1] 84d4c689-b437-41a5-bb62-d144a30c6219 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 957.993910] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-58bde7d8-2d96-4cb7-98a8-c5b20d28b6eb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.001848] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114415, 'name': Rename_Task, 'duration_secs': 0.172884} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.007808] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 958.008201] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 958.008201] env[65522]: value = "task-5114420" [ 958.008201] env[65522]: _type = "Task" [ 958.008201] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.008595] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-35c1edbf-5889-4aee-be19-dc4ddc2171f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.019990] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114420, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.021583] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 958.021583] env[65522]: value = "task-5114421" [ 958.021583] env[65522]: _type = "Task" [ 958.021583] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.040763] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114421, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.048030] env[65522]: WARNING openstack [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 958.048030] env[65522]: WARNING openstack [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 958.093070] env[65522]: DEBUG oslo_vmware.api [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114416, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.099504] env[65522]: WARNING neutronclient.v2_0.client [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 958.100201] env[65522]: WARNING openstack [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 958.100610] env[65522]: WARNING openstack [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 958.116291] env[65522]: DEBUG oslo_vmware.api [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114413, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.206196] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114417, 'name': CreateVM_Task, 'duration_secs': 0.4426} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.207525] env[65522]: WARNING neutronclient.v2_0.client [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 958.208192] env[65522]: WARNING openstack [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 958.208578] env[65522]: WARNING openstack [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 958.216751] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 958.218569] env[65522]: WARNING neutronclient.v2_0.client [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 958.218569] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.218569] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 958.218887] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 958.219106] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-728b23fb-747c-4278-b029-cd455cc61250 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.226193] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 958.226193] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fd4e1a-cae8-4661-075d-fe935e67dfe4" [ 958.226193] env[65522]: _type = "Task" [ 958.226193] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.237508] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fd4e1a-cae8-4661-075d-fe935e67dfe4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.285546] env[65522]: DEBUG nova.network.neutron [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Updated VIF entry in instance network info cache for port 626b09b1-0266-4a08-a37f-135470571842. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 958.285910] env[65522]: DEBUG nova.network.neutron [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Updating instance_info_cache with network_info: [{"id": "626b09b1-0266-4a08-a37f-135470571842", "address": "fa:16:3e:cb:66:49", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap626b09b1-02", "ovs_interfaceid": "626b09b1-0266-4a08-a37f-135470571842", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 958.436763] env[65522]: DEBUG nova.network.neutron [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating instance_info_cache with network_info: [{"id": "f8145f30-4e64-4a22-89a7-ffa393891ffb", "address": "fa:16:3e:42:70:f5", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8145f30-4e", "ovs_interfaceid": "f8145f30-4e64-4a22-89a7-ffa393891ffb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 958.464370] env[65522]: DEBUG oslo_vmware.api [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114414, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.522617] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114420, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234391} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.525933] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 958.526056] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 958.526136] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 958.544908] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114421, 'name': PowerOnVM_Task} progress is 71%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.601572] env[65522]: DEBUG oslo_vmware.api [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114416, 'name': PowerOffVM_Task, 'duration_secs': 0.92628} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.602230] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 958.602404] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 958.602679] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c6383dec-7ef0-45a3-aa32-24126143be89 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.610562] env[65522]: DEBUG oslo_vmware.api [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114413, 'name': ReconfigVM_Task, 'duration_secs': 1.187991} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.610859] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994827', 'volume_id': '33425e73-8f97-4928-afed-0733d11034fb', 'name': 'volume-33425e73-8f97-4928-afed-0733d11034fb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1eebb320-30e2-4d8b-b1fd-19e31fb35b77', 'attached_at': '', 'detached_at': '', 'volume_id': '33425e73-8f97-4928-afed-0733d11034fb', 'serial': '33425e73-8f97-4928-afed-0733d11034fb'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 958.614765] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d2a128-ebe8-4f4e-a4c8-ae88b1eff35b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.624168] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe8dd77-c65b-400d-8c1a-f1a8c603e7bf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.662370] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f71258-d98d-43b6-8f8c-24cb4bf7ee6a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.670708] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe89860-215f-4d9a-8687-7f5bace93c19 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.677534] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 958.677534] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 958.677534] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleting the datastore file [datastore2] 24f520ec-6a56-4f17-9ae4-2c856bf99582 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.678670] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d3de9e8d-dd23-4fab-89fd-89c59caf8183 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.691221] env[65522]: DEBUG nova.compute.provider_tree [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.699546] env[65522]: DEBUG oslo_vmware.api [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 958.699546] env[65522]: value = "task-5114423" [ 958.699546] env[65522]: _type = "Task" [ 958.699546] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.727759] env[65522]: DEBUG oslo_vmware.api [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114423, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.739631] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fd4e1a-cae8-4661-075d-fe935e67dfe4, 'name': SearchDatastore_Task, 'duration_secs': 0.022009} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.743028] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 958.743028] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 958.743028] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.743028] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 958.743028] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 958.743028] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be43f25b-859a-431e-94e2-d85ec7a6c839 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.752557] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 958.752805] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 958.753666] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bb43f73-0f00-46cc-b2df-e56cdfaa7405 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.764349] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 958.764349] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520ba8b4-49c0-7675-308e-86943cc89852" [ 958.764349] env[65522]: _type = "Task" [ 958.764349] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.774933] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520ba8b4-49c0-7675-308e-86943cc89852, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.789412] env[65522]: DEBUG oslo_concurrency.lockutils [req-1762a5ac-32c8-47dc-b7cf-987c9e062624 req-418fe5c9-2e41-4b97-95df-a797a47629ec service nova] Releasing lock "refresh_cache-1b2779f5-c6e8-4226-a819-0560d63bd7df" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 958.940608] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 958.959043] env[65522]: DEBUG oslo_vmware.api [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114414, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.043875] env[65522]: DEBUG oslo_vmware.api [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114421, 'name': PowerOnVM_Task, 'duration_secs': 0.817273} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.044179] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 959.044381] env[65522]: INFO nova.compute.manager [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Took 8.49 seconds to spawn the instance on the hypervisor. [ 959.044583] env[65522]: DEBUG nova.compute.manager [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 959.045421] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-950d030e-b7c2-4a7c-a3ee-2d84062311b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.195022] env[65522]: DEBUG nova.scheduler.client.report [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 959.213753] env[65522]: DEBUG oslo_vmware.api [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114423, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321347} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.214133] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.214242] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 959.214455] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 959.214874] env[65522]: INFO nova.compute.manager [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Took 1.65 seconds to destroy the instance on the hypervisor. [ 959.214874] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 959.215129] env[65522]: DEBUG nova.compute.manager [-] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 959.215221] env[65522]: DEBUG nova.network.neutron [-] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 959.215770] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 959.218527] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 959.218527] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 959.277445] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520ba8b4-49c0-7675-308e-86943cc89852, 'name': SearchDatastore_Task, 'duration_secs': 0.013679} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.278382] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49c29e37-cb32-4f6c-a26c-8b9873c25840 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.285536] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 959.285536] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52631cb6-c1dc-6540-4f2e-586199d8745f" [ 959.285536] env[65522]: _type = "Task" [ 959.285536] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.296079] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52631cb6-c1dc-6540-4f2e-586199d8745f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.303268] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 959.345639] env[65522]: DEBUG nova.objects.instance [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lazy-loading 'flavor' on Instance uuid 1eebb320-30e2-4d8b-b1fd-19e31fb35b77 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.466543] env[65522]: DEBUG oslo_vmware.api [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114414, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.479149] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6828236f-900e-4c28-bab4-ccbbc9d3eb34 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.499802] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7af8728-5049-4b9e-99c3-2b6bfead25c9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.508852] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating instance 'a2a5fb2d-62e0-4809-a01f-f1df66dca58f' progress to 83 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 959.575170] env[65522]: INFO nova.compute.manager [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Took 31.81 seconds to build instance. [ 959.581832] env[65522]: DEBUG nova.virt.hardware [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 959.582250] env[65522]: DEBUG nova.virt.hardware [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 959.582250] env[65522]: DEBUG nova.virt.hardware [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 959.582935] env[65522]: DEBUG nova.virt.hardware [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 959.583551] env[65522]: DEBUG nova.virt.hardware [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 959.583551] env[65522]: DEBUG nova.virt.hardware [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 959.583656] env[65522]: DEBUG nova.virt.hardware [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.583821] env[65522]: DEBUG nova.virt.hardware [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 959.583994] env[65522]: DEBUG nova.virt.hardware [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 959.584995] env[65522]: DEBUG nova.virt.hardware [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 959.584995] env[65522]: DEBUG nova.virt.hardware [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 959.585555] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba1f326a-2be6-43af-bbc0-a56352dcb195 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.596045] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c8d99e-9b63-47ae-8766-358ad1fc1ac1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.615261] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:f7:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56136ef6-99d7-4562-9a9f-d66fec951c5c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd06b43a-4102-4a7a-a82e-cf76b354ba13', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 959.622869] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 959.623497] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 959.623731] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea0eb58c-ea04-4f7d-a8ee-b55a0401342e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.642674] env[65522]: DEBUG nova.compute.manager [req-5720923c-923d-4a39-b487-9f6ba068b5eb req-2dd6704a-fedd-4157-9a97-a2153e10b5c8 service nova] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Received event network-vif-deleted-f33077f4-9852-477a-ae48-9e86ad80fdce {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 959.642831] env[65522]: INFO nova.compute.manager [req-5720923c-923d-4a39-b487-9f6ba068b5eb req-2dd6704a-fedd-4157-9a97-a2153e10b5c8 service nova] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Neutron deleted interface f33077f4-9852-477a-ae48-9e86ad80fdce; detaching it from the instance and deleting it from the info cache [ 959.642995] env[65522]: DEBUG nova.network.neutron [req-5720923c-923d-4a39-b487-9f6ba068b5eb req-2dd6704a-fedd-4157-9a97-a2153e10b5c8 service nova] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 959.649740] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 959.649740] env[65522]: value = "task-5114424" [ 959.649740] env[65522]: _type = "Task" [ 959.649740] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.659273] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114424, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.700991] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.509s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 959.701577] env[65522]: DEBUG nova.compute.manager [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 959.705063] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.376s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 959.705063] env[65522]: DEBUG nova.objects.instance [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Lazy-loading 'resources' on Instance uuid 21715aa5-24d4-423b-92a0-be7cc1f60877 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.797508] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52631cb6-c1dc-6540-4f2e-586199d8745f, 'name': SearchDatastore_Task, 'duration_secs': 0.015286} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.797792] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 959.798074] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 1b2779f5-c6e8-4226-a819-0560d63bd7df/1b2779f5-c6e8-4226-a819-0560d63bd7df.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 959.798354] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88553a15-77d9-4320-adab-79bb537e1f51 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.807631] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 959.807631] env[65522]: value = "task-5114425" [ 959.807631] env[65522]: _type = "Task" [ 959.807631] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.818444] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114425, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.964655] env[65522]: DEBUG oslo_vmware.api [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114414, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.016182] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 960.016182] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5076f12e-d772-418b-91d6-ad64c43d0541 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.026801] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 960.026801] env[65522]: value = "task-5114426" [ 960.026801] env[65522]: _type = "Task" [ 960.026801] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.041082] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114426, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.072722] env[65522]: DEBUG nova.network.neutron [-] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 960.076410] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3a4483c-6c13-4fbd-9190-9fedb803fcee tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.327s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 960.106125] env[65522]: DEBUG nova.compute.manager [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 960.108017] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab7d8ac9-24ed-41d8-b1c0-f3775fec9364 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.146483] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5ff60ba2-b6c4-4aaa-91ff-1ec9399b0602 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.163274] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-615630e2-83df-4ef1-9a7f-5c82f5f13394 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.216065] env[65522]: DEBUG nova.compute.utils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 960.222998] env[65522]: DEBUG nova.compute.manager [req-5720923c-923d-4a39-b487-9f6ba068b5eb req-2dd6704a-fedd-4157-9a97-a2153e10b5c8 service nova] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Detach interface failed, port_id=f33077f4-9852-477a-ae48-9e86ad80fdce, reason: Instance 24f520ec-6a56-4f17-9ae4-2c856bf99582 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 960.224733] env[65522]: DEBUG nova.compute.manager [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 960.224733] env[65522]: DEBUG nova.network.neutron [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 960.225179] env[65522]: WARNING neutronclient.v2_0.client [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 960.225244] env[65522]: WARNING neutronclient.v2_0.client [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 960.225945] env[65522]: WARNING openstack [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 960.226519] env[65522]: WARNING openstack [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 960.307412] env[65522]: DEBUG nova.policy [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8f8fa586d8c74bb9805a1148fa7ed037', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f93c790ff61543bd8e134bcf9cb20bb2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 960.328959] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114425, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.356443] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6d14f9d5-1a34-42e3-9ad6-136c361b96d1 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.482s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 960.374122] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114424, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.463603] env[65522]: DEBUG oslo_vmware.api [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114414, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.542573] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114426, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.576899] env[65522]: INFO nova.compute.manager [-] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Took 1.36 seconds to deallocate network for instance. [ 960.622085] env[65522]: INFO nova.compute.manager [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] instance snapshotting [ 960.622085] env[65522]: WARNING nova.compute.manager [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 960.624416] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15101e24-9d18-465a-aff5-899c11a034cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.650909] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf13503e-da22-4316-a032-5a6c979c8c84 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.654487] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a600741c-3733-48fe-8cff-8f375c1c93e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.658245] env[65522]: DEBUG nova.compute.manager [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Received event network-changed-bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 960.658442] env[65522]: DEBUG nova.compute.manager [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Refreshing instance network info cache due to event network-changed-bd432f78-055d-4cd1-b502-8e058700c03a. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 960.659507] env[65522]: DEBUG oslo_concurrency.lockutils [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] Acquiring lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.659507] env[65522]: DEBUG oslo_concurrency.lockutils [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] Acquired lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 960.659507] env[65522]: DEBUG nova.network.neutron [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Refreshing network info cache for port bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 960.674107] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac341699-ac0c-4a7b-b72c-62d0dc1f3dc1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.682511] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114424, 'name': CreateVM_Task, 'duration_secs': 0.52289} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.683437] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 960.683713] env[65522]: WARNING neutronclient.v2_0.client [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 960.684134] env[65522]: DEBUG oslo_concurrency.lockutils [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.684377] env[65522]: DEBUG oslo_concurrency.lockutils [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 960.684744] env[65522]: DEBUG oslo_concurrency.lockutils [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 960.685275] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7aba052b-7737-4e32-91eb-d6068e5515ca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.714815] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab89a80-85a4-47b0-ae7e-ead2cbb042d1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.720021] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 960.720021] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52167f59-15fb-a5fc-7ab1-b5f482949c41" [ 960.720021] env[65522]: _type = "Task" [ 960.720021] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.725212] env[65522]: DEBUG nova.compute.manager [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 960.729903] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e631af4-5d04-4c19-9cbb-b30bbc28628a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.738579] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52167f59-15fb-a5fc-7ab1-b5f482949c41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.750851] env[65522]: DEBUG nova.compute.provider_tree [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.758556] env[65522]: DEBUG nova.network.neutron [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Successfully created port: 37eeac6b-6857-4309-b861-5eb05cdba980 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 960.821892] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114425, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.709021} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.822204] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 1b2779f5-c6e8-4226-a819-0560d63bd7df/1b2779f5-c6e8-4226-a819-0560d63bd7df.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 960.822626] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.822702] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea6fcc1a-3264-4899-84b4-79ed0f383861 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.830452] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 960.830452] env[65522]: value = "task-5114427" [ 960.830452] env[65522]: _type = "Task" [ 960.830452] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.841038] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114427, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.965092] env[65522]: DEBUG oslo_vmware.api [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114414, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.042781] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114426, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.084269] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 961.114009] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 961.114357] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 961.114579] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 961.114762] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 961.114930] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 961.117326] env[65522]: INFO nova.compute.manager [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Terminating instance [ 961.164785] env[65522]: WARNING neutronclient.v2_0.client [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 961.164785] env[65522]: WARNING openstack [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.165137] env[65522]: WARNING openstack [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 961.188668] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 961.189028] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-954f16b1-ac27-492a-8df0-c79d1fecd361 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.197783] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 961.197783] env[65522]: value = "task-5114428" [ 961.197783] env[65522]: _type = "Task" [ 961.197783] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.207585] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114428, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.232883] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52167f59-15fb-a5fc-7ab1-b5f482949c41, 'name': SearchDatastore_Task, 'duration_secs': 0.047845} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.233332] env[65522]: DEBUG oslo_concurrency.lockutils [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 961.233659] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 961.233988] env[65522]: DEBUG oslo_concurrency.lockutils [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.234224] env[65522]: DEBUG oslo_concurrency.lockutils [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 961.234500] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 961.239187] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff198bd2-64fa-4c27-9f8f-be4cf2202ab4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.254951] env[65522]: DEBUG nova.scheduler.client.report [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 961.259785] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 961.259971] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 961.262132] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-082e1481-3f69-4388-bd4c-b85653013e17 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.269516] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 961.269516] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a65329-468c-9655-4e3c-2bfe6711dedb" [ 961.269516] env[65522]: _type = "Task" [ 961.269516] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.279358] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a65329-468c-9655-4e3c-2bfe6711dedb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.342128] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114427, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.134983} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.342128] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 961.342960] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65651a49-c2d0-4273-ae8e-da987ac89aed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.369754] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 1b2779f5-c6e8-4226-a819-0560d63bd7df/1b2779f5-c6e8-4226-a819-0560d63bd7df.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 961.370166] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc0c90dc-5faf-46f9-b0eb-128ca07bd37f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.395469] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 961.395469] env[65522]: value = "task-5114429" [ 961.395469] env[65522]: _type = "Task" [ 961.395469] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.408279] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114429, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.457986] env[65522]: WARNING openstack [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.458681] env[65522]: WARNING openstack [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 961.483279] env[65522]: DEBUG oslo_vmware.api [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114414, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.487724] env[65522]: DEBUG oslo_concurrency.lockutils [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "95a10a6f-8c6c-435d-a360-dff10eee0855" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 961.487926] env[65522]: DEBUG oslo_concurrency.lockutils [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "95a10a6f-8c6c-435d-a360-dff10eee0855" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 961.535933] env[65522]: WARNING neutronclient.v2_0.client [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 961.536127] env[65522]: WARNING openstack [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.536631] env[65522]: WARNING openstack [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 961.548576] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114426, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.621295] env[65522]: DEBUG nova.compute.manager [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 961.621740] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 961.623124] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516d41f5-d095-401d-8ca9-8cb39dc53803 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.636151] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 961.636460] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cfe112ca-acd9-4638-b425-e067bde8353e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.645512] env[65522]: DEBUG oslo_vmware.api [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 961.645512] env[65522]: value = "task-5114430" [ 961.645512] env[65522]: _type = "Task" [ 961.645512] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.656414] env[65522]: DEBUG oslo_vmware.api [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114430, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.694510] env[65522]: DEBUG nova.network.neutron [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Updated VIF entry in instance network info cache for port bd432f78-055d-4cd1-b502-8e058700c03a. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 961.694974] env[65522]: DEBUG nova.network.neutron [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Updating instance_info_cache with network_info: [{"id": "bd432f78-055d-4cd1-b502-8e058700c03a", "address": "fa:16:3e:38:8a:d1", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd432f78-05", "ovs_interfaceid": "bd432f78-055d-4cd1-b502-8e058700c03a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 961.708588] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114428, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.742407] env[65522]: DEBUG nova.compute.manager [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 961.770697] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.066s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 961.775035] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.426s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 961.775035] env[65522]: DEBUG nova.objects.instance [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Lazy-loading 'resources' on Instance uuid 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.783298] env[65522]: DEBUG nova.virt.hardware [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 961.784546] env[65522]: DEBUG nova.virt.hardware [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 961.784546] env[65522]: DEBUG nova.virt.hardware [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 961.784546] env[65522]: DEBUG nova.virt.hardware [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 961.784546] env[65522]: DEBUG nova.virt.hardware [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 961.784546] env[65522]: DEBUG nova.virt.hardware [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 961.784831] env[65522]: DEBUG nova.virt.hardware [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 961.784977] env[65522]: DEBUG nova.virt.hardware [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 961.785236] env[65522]: DEBUG nova.virt.hardware [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 961.785466] env[65522]: DEBUG nova.virt.hardware [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 961.785709] env[65522]: DEBUG nova.virt.hardware [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 961.788149] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63973000-b0bc-4f35-b24e-eba6c0778b65 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.803745] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68577895-acf0-441e-8c82-d2c8ad214ccf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.808312] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a65329-468c-9655-4e3c-2bfe6711dedb, 'name': SearchDatastore_Task, 'duration_secs': 0.016159} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.809647] env[65522]: INFO nova.scheduler.client.report [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Deleted allocations for instance 21715aa5-24d4-423b-92a0-be7cc1f60877 [ 961.821025] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9c3f778-d6c7-48de-9171-087b0a580e12 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.836962] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 961.836962] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52db1dfa-ff07-65ba-4706-9656dc973c77" [ 961.836962] env[65522]: _type = "Task" [ 961.836962] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.849839] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52db1dfa-ff07-65ba-4706-9656dc973c77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.906844] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114429, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.979530] env[65522]: DEBUG oslo_vmware.api [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114414, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.990193] env[65522]: DEBUG nova.compute.manager [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 962.040332] env[65522]: DEBUG oslo_vmware.api [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114426, 'name': PowerOnVM_Task, 'duration_secs': 1.749435} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.040697] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 962.040944] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8cacd363-3d48-4d2a-b2c2-35e8c9c62a8f tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating instance 'a2a5fb2d-62e0-4809-a01f-f1df66dca58f' progress to 100 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 962.115821] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "7701b92d-805e-4837-b900-326910b5eef5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 962.116012] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "7701b92d-805e-4837-b900-326910b5eef5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 962.157828] env[65522]: DEBUG oslo_vmware.api [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114430, 'name': PowerOffVM_Task, 'duration_secs': 0.243271} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.158122] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 962.158293] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 962.158556] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cef7feda-d6da-4490-9327-2412bb62c5ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.197867] env[65522]: DEBUG oslo_concurrency.lockutils [req-8d16ba8b-c03f-4c2b-aaae-bd2099b2f26d req-5e6df493-b44a-49b6-b17e-b4bc5e2408b5 service nova] Releasing lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 962.209453] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114428, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.233570] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 962.233740] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 962.233898] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Deleting the datastore file [datastore2] 1eebb320-30e2-4d8b-b1fd-19e31fb35b77 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 962.234132] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a48c63f1-a95f-4e50-8c85-8cfa16d2b90e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.241368] env[65522]: DEBUG oslo_vmware.api [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 962.241368] env[65522]: value = "task-5114432" [ 962.241368] env[65522]: _type = "Task" [ 962.241368] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.251876] env[65522]: DEBUG oslo_vmware.api [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114432, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.326603] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18680265-7865-4f36-8db3-5e6afe766881 tempest-ServerAddressesTestJSON-1052568870 tempest-ServerAddressesTestJSON-1052568870-project-member] Lock "21715aa5-24d4-423b-92a0-be7cc1f60877" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.926s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 962.352455] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52db1dfa-ff07-65ba-4706-9656dc973c77, 'name': SearchDatastore_Task, 'duration_secs': 0.012746} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.352732] env[65522]: DEBUG oslo_concurrency.lockutils [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 962.353048] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 84d4c689-b437-41a5-bb62-d144a30c6219/84d4c689-b437-41a5-bb62-d144a30c6219.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 962.353333] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7bf4e96f-c4e2-460d-bde4-21c3228c3ac6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.364139] env[65522]: DEBUG nova.network.neutron [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Successfully updated port: 37eeac6b-6857-4309-b861-5eb05cdba980 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 962.367366] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 962.367366] env[65522]: value = "task-5114433" [ 962.367366] env[65522]: _type = "Task" [ 962.367366] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.382799] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114433, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.408723] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114429, 'name': ReconfigVM_Task, 'duration_secs': 0.525978} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.411944] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 1b2779f5-c6e8-4226-a819-0560d63bd7df/1b2779f5-c6e8-4226-a819-0560d63bd7df.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 962.412862] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c27f3a49-01aa-4e47-b38f-9cdb6e6fd6ed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.420597] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 962.420597] env[65522]: value = "task-5114434" [ 962.420597] env[65522]: _type = "Task" [ 962.420597] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.431903] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114434, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.485911] env[65522]: DEBUG oslo_vmware.api [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114414, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.511114] env[65522]: DEBUG oslo_concurrency.lockutils [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 962.598067] env[65522]: DEBUG nova.compute.manager [req-8c7fde48-908b-4fa3-a7ab-a757c4d94602 req-81bc15b0-e953-4a07-a792-cb7c81391a33 service nova] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Received event network-vif-plugged-37eeac6b-6857-4309-b861-5eb05cdba980 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 962.598382] env[65522]: DEBUG oslo_concurrency.lockutils [req-8c7fde48-908b-4fa3-a7ab-a757c4d94602 req-81bc15b0-e953-4a07-a792-cb7c81391a33 service nova] Acquiring lock "82a7ad85-a061-4eec-bd3f-fc977532dfbf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 962.598658] env[65522]: DEBUG oslo_concurrency.lockutils [req-8c7fde48-908b-4fa3-a7ab-a757c4d94602 req-81bc15b0-e953-4a07-a792-cb7c81391a33 service nova] Lock "82a7ad85-a061-4eec-bd3f-fc977532dfbf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 962.598890] env[65522]: DEBUG oslo_concurrency.lockutils [req-8c7fde48-908b-4fa3-a7ab-a757c4d94602 req-81bc15b0-e953-4a07-a792-cb7c81391a33 service nova] Lock "82a7ad85-a061-4eec-bd3f-fc977532dfbf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 962.599114] env[65522]: DEBUG nova.compute.manager [req-8c7fde48-908b-4fa3-a7ab-a757c4d94602 req-81bc15b0-e953-4a07-a792-cb7c81391a33 service nova] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] No waiting events found dispatching network-vif-plugged-37eeac6b-6857-4309-b861-5eb05cdba980 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 962.599504] env[65522]: WARNING nova.compute.manager [req-8c7fde48-908b-4fa3-a7ab-a757c4d94602 req-81bc15b0-e953-4a07-a792-cb7c81391a33 service nova] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Received unexpected event network-vif-plugged-37eeac6b-6857-4309-b861-5eb05cdba980 for instance with vm_state building and task_state spawning. [ 962.617977] env[65522]: DEBUG nova.compute.manager [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 962.699918] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057b7bcb-4657-4f0d-8c75-6df09476d9e4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.718323] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-674dd67e-027e-4b52-aca2-153d81fa4782 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.721901] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114428, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.772372] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549d44c4-3640-4a25-b545-12a8aadca343 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.784265] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88109607-ba62-4667-9f5c-568ac431671c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.788814] env[65522]: DEBUG oslo_vmware.api [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114432, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.800985] env[65522]: DEBUG nova.compute.provider_tree [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.868952] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "refresh_cache-82a7ad85-a061-4eec-bd3f-fc977532dfbf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.869166] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "refresh_cache-82a7ad85-a061-4eec-bd3f-fc977532dfbf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 962.870346] env[65522]: DEBUG nova.network.neutron [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 962.883340] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114433, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.943692] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114434, 'name': Rename_Task, 'duration_secs': 0.167367} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.944358] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 962.944468] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-665cdd90-a8dd-478d-a3f4-c5340620f045 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.957287] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 962.957287] env[65522]: value = "task-5114435" [ 962.957287] env[65522]: _type = "Task" [ 962.957287] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.974330] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114435, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.987061] env[65522]: DEBUG oslo_vmware.api [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114414, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.146881] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 963.216353] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114428, 'name': CreateSnapshot_Task, 'duration_secs': 1.691028} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.216662] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 963.217554] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db2b1c5-0a7d-47be-a1a1-fc2c718b07dd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.282965] env[65522]: DEBUG oslo_vmware.api [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114432, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.662524} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.283246] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 963.283483] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 963.283627] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 963.283886] env[65522]: INFO nova.compute.manager [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Took 1.66 seconds to destroy the instance on the hypervisor. [ 963.284606] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 963.284606] env[65522]: DEBUG nova.compute.manager [-] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 963.284606] env[65522]: DEBUG nova.network.neutron [-] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 963.284857] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 963.285325] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 963.285828] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 963.304205] env[65522]: DEBUG nova.scheduler.client.report [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 963.349018] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 963.377669] env[65522]: WARNING openstack [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 963.379117] env[65522]: WARNING openstack [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 963.391358] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114433, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.934362} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.391761] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 84d4c689-b437-41a5-bb62-d144a30c6219/84d4c689-b437-41a5-bb62-d144a30c6219.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 963.392063] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 963.392435] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7923cf95-1969-4738-9070-c971c889a3b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.402343] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 963.402343] env[65522]: value = "task-5114436" [ 963.402343] env[65522]: _type = "Task" [ 963.402343] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.414439] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114436, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.471590] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114435, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.483656] env[65522]: DEBUG oslo_vmware.api [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114414, 'name': ReconfigVM_Task, 'duration_secs': 5.841371} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.483656] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 963.483839] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Reconfigured VM to detach interface {{(pid=65522) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 963.484328] env[65522]: WARNING neutronclient.v2_0.client [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 963.484695] env[65522]: WARNING neutronclient.v2_0.client [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 963.485429] env[65522]: WARNING openstack [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 963.485883] env[65522]: WARNING openstack [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 963.603787] env[65522]: DEBUG nova.network.neutron [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 963.719692] env[65522]: WARNING openstack [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 963.720245] env[65522]: WARNING openstack [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 963.741192] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 963.741589] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c0d10620-7ac3-4120-a8b0-53fee6c51c68 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.753056] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 963.753056] env[65522]: value = "task-5114437" [ 963.753056] env[65522]: _type = "Task" [ 963.753056] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.763589] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114437, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.809862] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.036s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 963.813270] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.132s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 963.814163] env[65522]: DEBUG nova.objects.instance [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lazy-loading 'resources' on Instance uuid 05319f69-2ac7-4649-a90a-3542d87a3e82 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.847194] env[65522]: INFO nova.scheduler.client.report [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Deleted allocations for instance 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2 [ 963.915844] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114436, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083756} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.916329] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 963.917855] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e98168-220a-42ff-9848-d2cade70fab1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.945117] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] 84d4c689-b437-41a5-bb62-d144a30c6219/84d4c689-b437-41a5-bb62-d144a30c6219.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.947598] env[65522]: WARNING neutronclient.v2_0.client [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 963.948207] env[65522]: WARNING openstack [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 963.948589] env[65522]: WARNING openstack [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 963.957036] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b9a6e0b-23e9-408a-b493-0193efc54717 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.983972] env[65522]: DEBUG oslo_vmware.api [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114435, 'name': PowerOnVM_Task, 'duration_secs': 0.756132} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.986044] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 963.986044] env[65522]: INFO nova.compute.manager [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Took 8.84 seconds to spawn the instance on the hypervisor. [ 963.986044] env[65522]: DEBUG nova.compute.manager [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 963.986247] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 963.986247] env[65522]: value = "task-5114438" [ 963.986247] env[65522]: _type = "Task" [ 963.986247] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.988072] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca22bc6-40eb-42a9-8fb6-11836b985792 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.007846] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.181373] env[65522]: DEBUG nova.network.neutron [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Updating instance_info_cache with network_info: [{"id": "37eeac6b-6857-4309-b861-5eb05cdba980", "address": "fa:16:3e:8d:5a:3a", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37eeac6b-68", "ovs_interfaceid": "37eeac6b-6857-4309-b861-5eb05cdba980", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 964.268402] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114437, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.303997] env[65522]: DEBUG nova.network.neutron [-] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 964.361712] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c811496-398f-4e13-83a6-b9b64bd8578e tempest-ServersAaction247Test-795004711 tempest-ServersAaction247Test-795004711-project-member] Lock "9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.357s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 964.415494] env[65522]: DEBUG nova.compute.manager [req-5b66aac6-dc20-4c6c-b7be-7ef0a27e3912 req-6259b9a0-419c-45f3-b636-65d6de0c2104 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Received event network-vif-deleted-d91694fd-ef8c-478e-8e0e-4e8ca56702a8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 964.415727] env[65522]: INFO nova.compute.manager [req-5b66aac6-dc20-4c6c-b7be-7ef0a27e3912 req-6259b9a0-419c-45f3-b636-65d6de0c2104 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Neutron deleted interface d91694fd-ef8c-478e-8e0e-4e8ca56702a8; detaching it from the instance and deleting it from the info cache [ 964.416029] env[65522]: DEBUG nova.network.neutron [req-5b66aac6-dc20-4c6c-b7be-7ef0a27e3912 req-6259b9a0-419c-45f3-b636-65d6de0c2104 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updating instance_info_cache with network_info: [{"id": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "address": "fa:16:3e:8e:34:0c", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6221fb0-8c", "ovs_interfaceid": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "04e1485f-54a3-42d2-bd3f-19ea0fa7babe", "address": "fa:16:3e:fc:58:37", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04e1485f-54", "ovs_interfaceid": "04e1485f-54a3-42d2-bd3f-19ea0fa7babe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 964.501184] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114438, 'name': ReconfigVM_Task, 'duration_secs': 0.411203} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.503815] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Reconfigured VM instance instance-0000004a to attach disk [datastore2] 84d4c689-b437-41a5-bb62-d144a30c6219/84d4c689-b437-41a5-bb62-d144a30c6219.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.505435] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1dc3c536-b8b0-4d18-a0f6-8f957b88f9c0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.514748] env[65522]: INFO nova.compute.manager [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Took 28.73 seconds to build instance. [ 964.524465] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 964.524465] env[65522]: value = "task-5114439" [ 964.524465] env[65522]: _type = "Task" [ 964.524465] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.535036] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114439, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.685426] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "refresh_cache-82a7ad85-a061-4eec-bd3f-fc977532dfbf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 964.685791] env[65522]: DEBUG nova.compute.manager [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Instance network_info: |[{"id": "37eeac6b-6857-4309-b861-5eb05cdba980", "address": "fa:16:3e:8d:5a:3a", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37eeac6b-68", "ovs_interfaceid": "37eeac6b-6857-4309-b861-5eb05cdba980", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 964.686557] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8d:5a:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37eeac6b-6857-4309-b861-5eb05cdba980', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 964.695192] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Creating folder: Project (f93c790ff61543bd8e134bcf9cb20bb2). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 964.695192] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1364d264-fac4-4952-ae62-009420c83d43 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.711261] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Created folder: Project (f93c790ff61543bd8e134bcf9cb20bb2) in parent group-v994660. [ 964.711499] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Creating folder: Instances. Parent ref: group-v994868. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 964.714537] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fd54624e-499f-40a4-8155-cf8c3416ef34 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.719900] env[65522]: DEBUG nova.compute.manager [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Received event network-changed-37eeac6b-6857-4309-b861-5eb05cdba980 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 964.721070] env[65522]: DEBUG nova.compute.manager [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Refreshing instance network info cache due to event network-changed-37eeac6b-6857-4309-b861-5eb05cdba980. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 964.721070] env[65522]: DEBUG oslo_concurrency.lockutils [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] Acquiring lock "refresh_cache-82a7ad85-a061-4eec-bd3f-fc977532dfbf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.721070] env[65522]: DEBUG oslo_concurrency.lockutils [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] Acquired lock "refresh_cache-82a7ad85-a061-4eec-bd3f-fc977532dfbf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 964.721070] env[65522]: DEBUG nova.network.neutron [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Refreshing network info cache for port 37eeac6b-6857-4309-b861-5eb05cdba980 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 964.729867] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e1259b-8a49-42d3-b955-a94bc5878871 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.734722] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Created folder: Instances in parent group-v994868. [ 964.734970] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 964.735620] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 964.736134] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f7592ee-a4d0-45db-861b-6a720ffb6de5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.755186] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14afe46-49bc-45ac-9d54-f4a7f2a2de45 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.763654] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 964.763654] env[65522]: value = "task-5114442" [ 964.763654] env[65522]: _type = "Task" [ 964.763654] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.796699] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114437, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.798484] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a45e6b-52b3-4ee2-9330-bd64d1bc0e14 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.804843] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114442, 'name': CreateVM_Task} progress is 15%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.809084] env[65522]: INFO nova.compute.manager [-] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Took 1.52 seconds to deallocate network for instance. [ 964.810837] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d55c5602-12dd-4c7b-ae0e-0260e6cfa541 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.829106] env[65522]: DEBUG nova.compute.provider_tree [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.923419] env[65522]: DEBUG oslo_concurrency.lockutils [req-5b66aac6-dc20-4c6c-b7be-7ef0a27e3912 req-6259b9a0-419c-45f3-b636-65d6de0c2104 service nova] Acquiring lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.923419] env[65522]: DEBUG oslo_concurrency.lockutils [req-5b66aac6-dc20-4c6c-b7be-7ef0a27e3912 req-6259b9a0-419c-45f3-b636-65d6de0c2104 service nova] Acquired lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 964.925874] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4ec4f3-d240-407f-afff-68f1a98488b6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.948605] env[65522]: DEBUG oslo_concurrency.lockutils [req-5b66aac6-dc20-4c6c-b7be-7ef0a27e3912 req-6259b9a0-419c-45f3-b636-65d6de0c2104 service nova] Releasing lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 964.949169] env[65522]: WARNING nova.compute.manager [req-5b66aac6-dc20-4c6c-b7be-7ef0a27e3912 req-6259b9a0-419c-45f3-b636-65d6de0c2104 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Detach interface failed, port_id=d91694fd-ef8c-478e-8e0e-4e8ca56702a8, reason: No device with interface-id d91694fd-ef8c-478e-8e0e-4e8ca56702a8 exists on VM: nova.exception.NotFound: No device with interface-id d91694fd-ef8c-478e-8e0e-4e8ca56702a8 exists on VM [ 965.019507] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a12a6a3e-06cb-4527-b8e0-450e7170a7b5 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "1b2779f5-c6e8-4226-a819-0560d63bd7df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.248s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 965.022287] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.022448] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 965.022667] env[65522]: DEBUG nova.network.neutron [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 965.035380] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114439, 'name': Rename_Task, 'duration_secs': 0.178212} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.036371] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 965.036629] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93c19033-76b6-410c-b9b8-92aa326d600f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.045290] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 965.045290] env[65522]: value = "task-5114443" [ 965.045290] env[65522]: _type = "Task" [ 965.045290] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.054666] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114443, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.058434] env[65522]: DEBUG oslo_concurrency.lockutils [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 965.059855] env[65522]: DEBUG oslo_concurrency.lockutils [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 965.059855] env[65522]: DEBUG nova.compute.manager [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Going to confirm migration 4 {{(pid=65522) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 965.082331] env[65522]: DEBUG oslo_concurrency.lockutils [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 965.082796] env[65522]: DEBUG oslo_concurrency.lockutils [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 965.083106] env[65522]: DEBUG oslo_concurrency.lockutils [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 965.083297] env[65522]: DEBUG oslo_concurrency.lockutils [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 965.083809] env[65522]: DEBUG oslo_concurrency.lockutils [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 965.087400] env[65522]: INFO nova.compute.manager [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Terminating instance [ 965.224591] env[65522]: WARNING neutronclient.v2_0.client [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 965.225682] env[65522]: WARNING openstack [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 965.226253] env[65522]: WARNING openstack [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 965.268213] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114437, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.278519] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114442, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.333071] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 965.334138] env[65522]: DEBUG nova.scheduler.client.report [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 965.422594] env[65522]: WARNING openstack [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 965.423094] env[65522]: WARNING openstack [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 965.514658] env[65522]: WARNING neutronclient.v2_0.client [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 965.515498] env[65522]: WARNING openstack [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 965.515919] env[65522]: WARNING openstack [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 965.525450] env[65522]: WARNING neutronclient.v2_0.client [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 965.526338] env[65522]: WARNING openstack [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 965.526819] env[65522]: WARNING openstack [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 965.556869] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114443, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.564892] env[65522]: WARNING neutronclient.v2_0.client [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 965.591701] env[65522]: DEBUG nova.compute.manager [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 965.591986] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 965.593017] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c954c434-33ec-49c3-ae03-c7c05188353f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.603269] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 965.604049] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80ca84ed-9759-42dd-84a2-ad936a99f9a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.614594] env[65522]: DEBUG oslo_vmware.api [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 965.614594] env[65522]: value = "task-5114444" [ 965.614594] env[65522]: _type = "Task" [ 965.614594] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.617411] env[65522]: WARNING neutronclient.v2_0.client [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 965.617779] env[65522]: DEBUG oslo_concurrency.lockutils [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.617962] env[65522]: DEBUG oslo_concurrency.lockutils [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 965.618223] env[65522]: DEBUG nova.network.neutron [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 965.618416] env[65522]: DEBUG nova.objects.instance [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lazy-loading 'info_cache' on Instance uuid a2a5fb2d-62e0-4809-a01f-f1df66dca58f {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.626151] env[65522]: DEBUG oslo_vmware.api [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114444, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.683482] env[65522]: DEBUG nova.network.neutron [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Updated VIF entry in instance network info cache for port 37eeac6b-6857-4309-b861-5eb05cdba980. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 965.683914] env[65522]: DEBUG nova.network.neutron [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Updating instance_info_cache with network_info: [{"id": "37eeac6b-6857-4309-b861-5eb05cdba980", "address": "fa:16:3e:8d:5a:3a", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37eeac6b-68", "ovs_interfaceid": "37eeac6b-6857-4309-b861-5eb05cdba980", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 965.700324] env[65522]: WARNING openstack [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 965.700971] env[65522]: WARNING openstack [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 965.765863] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114437, 'name': CloneVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.780602] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114442, 'name': CreateVM_Task, 'duration_secs': 0.592548} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.780801] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 965.781242] env[65522]: WARNING neutronclient.v2_0.client [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 965.781856] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.782343] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 965.783079] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 965.783634] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca33585e-83d0-4632-9d6a-acfed2f419f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.791656] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 965.791656] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d41647-750a-1030-c885-825a7317c163" [ 965.791656] env[65522]: _type = "Task" [ 965.791656] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.801689] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d41647-750a-1030-c885-825a7317c163, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.803161] env[65522]: WARNING neutronclient.v2_0.client [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 965.803828] env[65522]: WARNING openstack [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 965.804263] env[65522]: WARNING openstack [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 965.842719] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.029s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 965.846205] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.681s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 965.846531] env[65522]: DEBUG nova.objects.instance [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lazy-loading 'resources' on Instance uuid d5e8dd05-dc3c-4831-b4b0-ac100360f3e7 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.878752] env[65522]: INFO nova.scheduler.client.report [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Deleted allocations for instance 05319f69-2ac7-4649-a90a-3542d87a3e82 [ 965.898406] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "1b2779f5-c6e8-4226-a819-0560d63bd7df" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 965.898699] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "1b2779f5-c6e8-4226-a819-0560d63bd7df" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 965.898921] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "1b2779f5-c6e8-4226-a819-0560d63bd7df-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 965.899123] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "1b2779f5-c6e8-4226-a819-0560d63bd7df-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 965.899389] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "1b2779f5-c6e8-4226-a819-0560d63bd7df-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 965.901790] env[65522]: INFO nova.compute.manager [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Terminating instance [ 965.916973] env[65522]: INFO nova.network.neutron [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Port 04e1485f-54a3-42d2-bd3f-19ea0fa7babe from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 965.917367] env[65522]: DEBUG nova.network.neutron [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updating instance_info_cache with network_info: [{"id": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "address": "fa:16:3e:8e:34:0c", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6221fb0-8c", "ovs_interfaceid": "c6221fb0-8c5b-478e-937e-3997556d0aaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 966.059709] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114443, 'name': PowerOnVM_Task, 'duration_secs': 0.553586} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.060118] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 966.060420] env[65522]: DEBUG nova.compute.manager [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 966.061575] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556e8fa8-fa0e-4f8e-901b-01a479bb7769 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.124769] env[65522]: DEBUG oslo_vmware.api [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114444, 'name': PowerOffVM_Task, 'duration_secs': 0.286739} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.125342] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 966.125514] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 966.125774] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c5eb0f89-11d0-481b-bca7-57428896db1a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.186972] env[65522]: DEBUG oslo_concurrency.lockutils [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] Releasing lock "refresh_cache-82a7ad85-a061-4eec-bd3f-fc977532dfbf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 966.187471] env[65522]: DEBUG nova.compute.manager [req-52aa9992-9585-4f56-be63-184f3637be6d req-94308c43-a19f-4bd8-8b75-08c78f261efa service nova] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Received event network-vif-deleted-e12c2307-4963-4a5a-af0f-19369217f430 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 966.250261] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 966.250261] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 966.250261] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Deleting the datastore file [datastore1] 9df1a51a-2811-4486-a4c6-58d618f2ae7d {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 966.250576] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-52803856-d90f-40a2-963b-488b5c749f60 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.260344] env[65522]: DEBUG oslo_vmware.api [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 966.260344] env[65522]: value = "task-5114446" [ 966.260344] env[65522]: _type = "Task" [ 966.260344] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.270056] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114437, 'name': CloneVM_Task, 'duration_secs': 2.105261} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.270736] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Created linked-clone VM from snapshot [ 966.272944] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f2df61-0de1-4496-996c-548b368551e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.279124] env[65522]: DEBUG oslo_vmware.api [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114446, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.285722] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Uploading image 57ba37b7-4542-434d-8a76-5c4f6d75462e {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 966.303638] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d41647-750a-1030-c885-825a7317c163, 'name': SearchDatastore_Task, 'duration_secs': 0.013635} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.303878] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 966.304143] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 966.304394] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.304542] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 966.304710] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.305030] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-912f2f0f-c8cf-48aa-bc1d-fad85f4ad886 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.320432] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 966.320432] env[65522]: value = "vm-994867" [ 966.320432] env[65522]: _type = "VirtualMachine" [ 966.320432] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 966.321582] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ebb68588-ca93-4685-946b-6a1566746ed9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.324916] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 966.325163] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 966.326698] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b199ac6-8673-452c-8340-9b595e74d7c4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.335543] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lease: (returnval){ [ 966.335543] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f27dba-2c2d-7656-72c9-f4a4ab4d7b92" [ 966.335543] env[65522]: _type = "HttpNfcLease" [ 966.335543] env[65522]: } obtained for exporting VM: (result){ [ 966.335543] env[65522]: value = "vm-994867" [ 966.335543] env[65522]: _type = "VirtualMachine" [ 966.335543] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 966.336649] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the lease: (returnval){ [ 966.336649] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f27dba-2c2d-7656-72c9-f4a4ab4d7b92" [ 966.336649] env[65522]: _type = "HttpNfcLease" [ 966.336649] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 966.337773] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 966.337773] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526a3aa1-724d-85c9-3a8b-435bd48ed18c" [ 966.337773] env[65522]: _type = "Task" [ 966.337773] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.352938] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 966.352938] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f27dba-2c2d-7656-72c9-f4a4ab4d7b92" [ 966.352938] env[65522]: _type = "HttpNfcLease" [ 966.352938] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 966.358123] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526a3aa1-724d-85c9-3a8b-435bd48ed18c, 'name': SearchDatastore_Task, 'duration_secs': 0.015454} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.359991] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c024ed8f-c6ae-4346-8811-d5d7d9ae37d5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.366356] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 966.366356] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f86e8d-b6a3-5a8b-ad57-eb432ed47065" [ 966.366356] env[65522]: _type = "Task" [ 966.366356] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.377563] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f86e8d-b6a3-5a8b-ad57-eb432ed47065, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.387972] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2ffa278e-769a-428d-bc2e-cb35eb75c576 tempest-ServersWithSpecificFlavorTestJSON-565045701 tempest-ServersWithSpecificFlavorTestJSON-565045701-project-member] Lock "05319f69-2ac7-4649-a90a-3542d87a3e82" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.489s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 966.406535] env[65522]: DEBUG nova.compute.manager [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 966.406781] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 966.410912] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c8111c3-dd76-4a30-8f9b-7c3cd7c72e81 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.420997] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "refresh_cache-9df1a51a-2811-4486-a4c6-58d618f2ae7d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 966.423886] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 966.424416] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10b8081b-d9ca-4844-981e-0807d284b319 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.436183] env[65522]: DEBUG oslo_vmware.api [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 966.436183] env[65522]: value = "task-5114448" [ 966.436183] env[65522]: _type = "Task" [ 966.436183] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.451190] env[65522]: DEBUG oslo_vmware.api [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114448, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.510701] env[65522]: DEBUG nova.compute.manager [req-e17ae83a-a3e3-4fd4-8bed-7f1db47498d0 req-c7a1826f-7d4f-4d5e-a69b-cc4fb02bafb9 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Received event network-vif-deleted-04e1485f-54a3-42d2-bd3f-19ea0fa7babe {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 966.574170] env[65522]: INFO nova.compute.manager [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] bringing vm to original state: 'stopped' [ 966.628669] env[65522]: WARNING neutronclient.v2_0.client [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 966.629711] env[65522]: WARNING openstack [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 966.629711] env[65522]: WARNING openstack [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 966.719087] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a5d68c-9126-44f1-bcff-8994d7aa3b93 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.730784] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e67add-59fc-4e56-bb27-0284812ee8bd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.767251] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b6408d8-dadb-48d7-b5ae-295852fe9e86 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.776260] env[65522]: DEBUG oslo_vmware.api [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114446, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.349718} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.778991] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.778991] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 966.779143] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 966.779378] env[65522]: INFO nova.compute.manager [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Took 1.19 seconds to destroy the instance on the hypervisor. [ 966.782564] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 966.782564] env[65522]: DEBUG nova.compute.manager [-] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 966.782564] env[65522]: DEBUG nova.network.neutron [-] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 966.782564] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 966.782564] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 966.782564] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 966.790552] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e64b6fc-f43b-4d61-8e46-c773ac6dc12c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.807149] env[65522]: DEBUG nova.compute.provider_tree [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.847170] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 966.847170] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f27dba-2c2d-7656-72c9-f4a4ab4d7b92" [ 966.847170] env[65522]: _type = "HttpNfcLease" [ 966.847170] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 966.847170] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 966.847170] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f27dba-2c2d-7656-72c9-f4a4ab4d7b92" [ 966.847170] env[65522]: _type = "HttpNfcLease" [ 966.847170] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 966.847170] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05214a69-7d47-41d8-83f1-f3c29dd1e792 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.855164] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52798a94-6485-7710-c375-de9f93b9b1ea/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 966.855346] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52798a94-6485-7710-c375-de9f93b9b1ea/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 966.920125] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f86e8d-b6a3-5a8b-ad57-eb432ed47065, 'name': SearchDatastore_Task, 'duration_secs': 0.028041} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.920405] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 966.920680] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 82a7ad85-a061-4eec-bd3f-fc977532dfbf/82a7ad85-a061-4eec-bd3f-fc977532dfbf.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 966.920985] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-59895083-8a5a-46ed-96b8-7094abddbbb1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.925116] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db0eeac3-94fb-43a3-8758-3b925b4b3ab7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-9df1a51a-2811-4486-a4c6-58d618f2ae7d-d91694fd-ef8c-478e-8e0e-4e8ca56702a8" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.064s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 966.932603] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 966.932603] env[65522]: value = "task-5114449" [ 966.932603] env[65522]: _type = "Task" [ 966.932603] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.943014] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114449, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.949963] env[65522]: DEBUG oslo_vmware.api [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114448, 'name': PowerOffVM_Task, 'duration_secs': 0.223528} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.950301] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 966.950490] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 966.950783] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ddcae03b-4772-4ac3-b5e3-16dcdcb25f0f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.974436] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 967.003988] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e63d2a22-d31f-4f58-bd2a-e48448a530b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.023299] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 967.023540] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 967.024153] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleting the datastore file [datastore1] 1b2779f5-c6e8-4226-a819-0560d63bd7df {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 967.024153] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-845303f0-61f2-4ba2-9e1e-a621849cabd3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.034809] env[65522]: DEBUG oslo_vmware.api [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 967.034809] env[65522]: value = "task-5114451" [ 967.034809] env[65522]: _type = "Task" [ 967.034809] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.052870] env[65522]: DEBUG oslo_vmware.api [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114451, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.054462] env[65522]: WARNING openstack [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 967.054895] env[65522]: WARNING openstack [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 967.183337] env[65522]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 04e1485f-54a3-42d2-bd3f-19ea0fa7babe could not be found.", "detail": ""}} {{(pid=65522) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:265}} [ 967.183616] env[65522]: DEBUG nova.network.neutron [-] Unable to show port 04e1485f-54a3-42d2-bd3f-19ea0fa7babe as it no longer exists. {{(pid=65522) _unbind_ports /opt/stack/nova/nova/network/neutron.py:700}} [ 967.309771] env[65522]: DEBUG nova.scheduler.client.report [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 967.446130] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114449, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.552814] env[65522]: DEBUG oslo_vmware.api [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114451, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.374776} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.554186] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 967.554497] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 967.554960] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 967.554960] env[65522]: INFO nova.compute.manager [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Took 1.15 seconds to destroy the instance on the hypervisor. [ 967.555538] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 967.555635] env[65522]: DEBUG nova.compute.manager [-] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 967.555722] env[65522]: DEBUG nova.network.neutron [-] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 967.557030] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 967.557030] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 967.557030] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 967.588159] env[65522]: DEBUG oslo_concurrency.lockutils [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "84d4c689-b437-41a5-bb62-d144a30c6219" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 967.588159] env[65522]: DEBUG oslo_concurrency.lockutils [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "84d4c689-b437-41a5-bb62-d144a30c6219" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 967.588508] env[65522]: DEBUG nova.compute.manager [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 967.589702] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d407ac9-a5a4-462e-bb6f-58feab640285 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.602365] env[65522]: DEBUG nova.compute.manager [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65522) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 967.804305] env[65522]: WARNING neutronclient.v2_0.client [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 967.804970] env[65522]: WARNING openstack [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 967.805332] env[65522]: WARNING openstack [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 967.817424] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.971s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.820422] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.736s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 967.820855] env[65522]: DEBUG nova.objects.instance [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lazy-loading 'resources' on Instance uuid 24f520ec-6a56-4f17-9ae4-2c856bf99582 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.696027] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 968.698226] env[65522]: INFO nova.scheduler.client.report [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Deleted allocations for instance d5e8dd05-dc3c-4831-b4b0-ac100360f3e7 [ 968.699903] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.700169] env[65522]: DEBUG nova.network.neutron [-] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 968.704846] env[65522]: DEBUG nova.compute.manager [req-f70dd6ec-54a2-4d6a-ba18-6ff48d36737a req-8a67a6e5-c512-4b9d-b9f1-e527679a3d96 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Received event network-vif-deleted-c6221fb0-8c5b-478e-937e-3997556d0aaa {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 968.705050] env[65522]: INFO nova.compute.manager [req-f70dd6ec-54a2-4d6a-ba18-6ff48d36737a req-8a67a6e5-c512-4b9d-b9f1-e527679a3d96 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Neutron deleted interface c6221fb0-8c5b-478e-937e-3997556d0aaa; detaching it from the instance and deleting it from the info cache [ 968.705377] env[65522]: DEBUG nova.network.neutron [req-f70dd6ec-54a2-4d6a-ba18-6ff48d36737a req-8a67a6e5-c512-4b9d-b9f1-e527679a3d96 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 968.713429] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5889c77-7753-4424-a663-45b99c76d8fa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.723813] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114449, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.810157} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.726170] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 82a7ad85-a061-4eec-bd3f-fc977532dfbf/82a7ad85-a061-4eec-bd3f-fc977532dfbf.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 968.726433] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 968.730041] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 968.730041] env[65522]: value = "task-5114452" [ 968.730041] env[65522]: _type = "Task" [ 968.730041] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.730041] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be8aaae3-7a8e-45d1-86a6-f23bd294d4c2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.746217] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114452, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.751948] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 968.751948] env[65522]: value = "task-5114453" [ 968.751948] env[65522]: _type = "Task" [ 968.751948] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.763921] env[65522]: DEBUG nova.network.neutron [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating instance_info_cache with network_info: [{"id": "f8145f30-4e64-4a22-89a7-ffa393891ffb", "address": "fa:16:3e:42:70:f5", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8145f30-4e", "ovs_interfaceid": "f8145f30-4e64-4a22-89a7-ffa393891ffb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 968.768988] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114453, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.127885] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c3a6f4e-620b-4fe8-87c6-b258e3f5d1dd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.137496] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b2b0791-ab12-4e60-94db-b0795f426146 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.179125] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3edfe50d-a28a-4b85-800c-7a7913b8317a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.188659] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8551049-4467-43ca-9e3f-95274ed8e892 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.208811] env[65522]: INFO nova.compute.manager [-] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Took 2.43 seconds to deallocate network for instance. [ 969.209528] env[65522]: DEBUG nova.compute.provider_tree [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.222553] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-88ac188d-6d26-4af6-9b71-bfaf022314a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.226116] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82a88afc-cef2-44bc-ba21-3caa969bcf91 tempest-MigrationsAdminTest-1397841056 tempest-MigrationsAdminTest-1397841056-project-member] Lock "d5e8dd05-dc3c-4831-b4b0-ac100360f3e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.044s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 969.240990] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5957c634-fedf-4c28-afdf-5ebc26f04b9e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.262317] env[65522]: DEBUG oslo_vmware.api [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114452, 'name': PowerOffVM_Task, 'duration_secs': 0.227879} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.266059] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.266338] env[65522]: DEBUG nova.compute.manager [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 969.267559] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a56e74f-9aeb-463a-b4c2-b1d3a4895407 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.276524] env[65522]: DEBUG oslo_concurrency.lockutils [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "refresh_cache-a2a5fb2d-62e0-4809-a01f-f1df66dca58f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 969.276738] env[65522]: DEBUG nova.objects.instance [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lazy-loading 'migration_context' on Instance uuid a2a5fb2d-62e0-4809-a01f-f1df66dca58f {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.278544] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114453, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.120245} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.282020] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 969.302011] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f26f64d-44d0-4f7d-a63e-c92854aac487 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.308295] env[65522]: DEBUG nova.compute.manager [req-f70dd6ec-54a2-4d6a-ba18-6ff48d36737a req-8a67a6e5-c512-4b9d-b9f1-e527679a3d96 service nova] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Detach interface failed, port_id=c6221fb0-8c5b-478e-937e-3997556d0aaa, reason: Instance 9df1a51a-2811-4486-a4c6-58d618f2ae7d could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 969.332232] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 82a7ad85-a061-4eec-bd3f-fc977532dfbf/82a7ad85-a061-4eec-bd3f-fc977532dfbf.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 969.332802] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d62b9e3-ee02-4b40-93eb-9e4bbf3b2e65 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.360268] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 969.360268] env[65522]: value = "task-5114454" [ 969.360268] env[65522]: _type = "Task" [ 969.360268] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.372259] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114454, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.549740] env[65522]: DEBUG nova.network.neutron [-] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 969.718727] env[65522]: DEBUG nova.scheduler.client.report [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 969.729876] env[65522]: DEBUG oslo_concurrency.lockutils [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 969.782731] env[65522]: DEBUG nova.objects.base [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 969.783767] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b23efce6-f0be-4395-8a76-196a66625548 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.811533] env[65522]: DEBUG oslo_concurrency.lockutils [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "84d4c689-b437-41a5-bb62-d144a30c6219" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.223s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 969.813768] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18c87cf4-6640-4aed-b3a3-8e71e5168445 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.821496] env[65522]: DEBUG oslo_vmware.api [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 969.821496] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e59b00-8073-3c78-e688-010f9edbba7c" [ 969.821496] env[65522]: _type = "Task" [ 969.821496] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.832917] env[65522]: DEBUG oslo_vmware.api [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e59b00-8073-3c78-e688-010f9edbba7c, 'name': SearchDatastore_Task} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.833278] env[65522]: DEBUG oslo_concurrency.lockutils [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 969.871356] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114454, 'name': ReconfigVM_Task, 'duration_secs': 0.382147} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.872323] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 82a7ad85-a061-4eec-bd3f-fc977532dfbf/82a7ad85-a061-4eec-bd3f-fc977532dfbf.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.872481] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9324ffef-b216-4418-8002-d6ffdc18422d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.881832] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 969.881832] env[65522]: value = "task-5114455" [ 969.881832] env[65522]: _type = "Task" [ 969.881832] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.892780] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114455, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.060268] env[65522]: INFO nova.compute.manager [-] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Took 2.50 seconds to deallocate network for instance. [ 970.226083] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.406s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 970.231459] env[65522]: DEBUG oslo_concurrency.lockutils [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.720s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.232492] env[65522]: INFO nova.compute.claims [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 970.273979] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Acquiring lock "00fcbbcb-ef87-4318-8c6e-ce62feb9fd83" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 970.274428] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Lock "00fcbbcb-ef87-4318-8c6e-ce62feb9fd83" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.276369] env[65522]: INFO nova.scheduler.client.report [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleted allocations for instance 24f520ec-6a56-4f17-9ae4-2c856bf99582 [ 970.321771] env[65522]: DEBUG oslo_concurrency.lockutils [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 970.394386] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114455, 'name': Rename_Task, 'duration_secs': 0.178007} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.394720] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 970.395030] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85f9abfc-99e0-4cee-b56b-06e92ce68517 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.403203] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 970.403203] env[65522]: value = "task-5114456" [ 970.403203] env[65522]: _type = "Task" [ 970.403203] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.413330] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114456, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.424567] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Acquiring lock "68d8d59b-4fc8-4993-8c9e-31b2329c88dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 970.424961] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Lock "68d8d59b-4fc8-4993-8c9e-31b2329c88dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.425225] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Acquiring lock "68d8d59b-4fc8-4993-8c9e-31b2329c88dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 970.428976] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Lock "68d8d59b-4fc8-4993-8c9e-31b2329c88dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.428976] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Lock "68d8d59b-4fc8-4993-8c9e-31b2329c88dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 970.428976] env[65522]: INFO nova.compute.manager [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Terminating instance [ 970.454625] env[65522]: DEBUG nova.compute.manager [req-ec52becd-8c01-4a2a-8c8b-d879d2c389b1 req-68498145-f350-4cf4-8988-f72d9f804f6d service nova] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Received event network-vif-deleted-626b09b1-0266-4a08-a37f-135470571842 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 970.571219] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 970.779251] env[65522]: DEBUG nova.compute.manager [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 970.787563] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a222450f-eb3c-4f8b-a696-c3e86b14922a tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "24f520ec-6a56-4f17-9ae4-2c856bf99582" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.746s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 970.919974] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114456, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.933265] env[65522]: DEBUG nova.compute.manager [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 970.933711] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 970.936093] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4fdf13-12ef-4856-93b1-af6f482e8206 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.948364] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.949491] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48c30236-17f2-48b9-9e0b-3c6e40fe02ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.959275] env[65522]: DEBUG oslo_vmware.api [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Waiting for the task: (returnval){ [ 970.959275] env[65522]: value = "task-5114457" [ 970.959275] env[65522]: _type = "Task" [ 970.959275] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.971056] env[65522]: DEBUG oslo_vmware.api [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5114457, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.310428] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.313467] env[65522]: DEBUG oslo_concurrency.lockutils [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "84d4c689-b437-41a5-bb62-d144a30c6219" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.313707] env[65522]: DEBUG oslo_concurrency.lockutils [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "84d4c689-b437-41a5-bb62-d144a30c6219" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 971.313926] env[65522]: DEBUG oslo_concurrency.lockutils [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "84d4c689-b437-41a5-bb62-d144a30c6219-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.314217] env[65522]: DEBUG oslo_concurrency.lockutils [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "84d4c689-b437-41a5-bb62-d144a30c6219-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 971.314300] env[65522]: DEBUG oslo_concurrency.lockutils [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "84d4c689-b437-41a5-bb62-d144a30c6219-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 971.316847] env[65522]: INFO nova.compute.manager [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Terminating instance [ 971.418897] env[65522]: DEBUG oslo_vmware.api [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114456, 'name': PowerOnVM_Task, 'duration_secs': 0.517112} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.421840] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 971.422067] env[65522]: INFO nova.compute.manager [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Took 9.68 seconds to spawn the instance on the hypervisor. [ 971.422243] env[65522]: DEBUG nova.compute.manager [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 971.423281] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f7d5ca0-eaa3-401b-b3b3-736f8a21801f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.475908] env[65522]: DEBUG oslo_vmware.api [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5114457, 'name': PowerOffVM_Task, 'duration_secs': 0.338194} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.477028] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 971.477028] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 971.477028] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4a23fc7-1148-4b36-a19b-f7595f27659b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.557668] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 971.557980] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 971.558270] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Deleting the datastore file [datastore2] 68d8d59b-4fc8-4993-8c9e-31b2329c88dd {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.558575] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5db8a4ba-eb5f-4491-a7f8-16ec09a2ea67 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.567015] env[65522]: DEBUG oslo_vmware.api [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Waiting for the task: (returnval){ [ 971.567015] env[65522]: value = "task-5114459" [ 971.567015] env[65522]: _type = "Task" [ 971.567015] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.583485] env[65522]: DEBUG oslo_vmware.api [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5114459, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.586896] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14b2e7f-1825-48c9-886b-6581c9a3e8c8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.594988] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4759c83b-2c9e-4b7c-834d-d3ed45ab86d5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.627079] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a87eaf4-4d51-419a-a1cf-76269cf0c789 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.635889] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12232430-b0a6-4e3d-80f1-7e9e2997ce5f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.651180] env[65522]: DEBUG nova.compute.provider_tree [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.775099] env[65522]: DEBUG oslo_concurrency.lockutils [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "bb20d99f-fc49-47bf-a6b7-74ea966b71eb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.775419] env[65522]: DEBUG oslo_concurrency.lockutils [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "bb20d99f-fc49-47bf-a6b7-74ea966b71eb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 971.775670] env[65522]: DEBUG oslo_concurrency.lockutils [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "bb20d99f-fc49-47bf-a6b7-74ea966b71eb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.775876] env[65522]: DEBUG oslo_concurrency.lockutils [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "bb20d99f-fc49-47bf-a6b7-74ea966b71eb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 971.776057] env[65522]: DEBUG oslo_concurrency.lockutils [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "bb20d99f-fc49-47bf-a6b7-74ea966b71eb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 971.781022] env[65522]: INFO nova.compute.manager [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Terminating instance [ 971.824182] env[65522]: DEBUG nova.compute.manager [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 971.824182] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 971.824182] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdbacd05-f7ff-4dd9-81ac-8155bde8e78e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.832128] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 971.832382] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c527518e-e44c-476c-ada7-b213d3143707 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.908931] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 971.909279] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 971.909546] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleting the datastore file [datastore2] 84d4c689-b437-41a5-bb62-d144a30c6219 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.909965] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f760b828-1bb5-4cea-8e68-30238b9448e3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.918433] env[65522]: DEBUG oslo_vmware.api [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 971.918433] env[65522]: value = "task-5114461" [ 971.918433] env[65522]: _type = "Task" [ 971.918433] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.927964] env[65522]: DEBUG oslo_vmware.api [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114461, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.947060] env[65522]: INFO nova.compute.manager [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Took 27.31 seconds to build instance. [ 972.085440] env[65522]: DEBUG oslo_vmware.api [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Task: {'id': task-5114459, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.334926} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.085728] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.087216] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 972.087216] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 972.087216] env[65522]: INFO nova.compute.manager [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Took 1.15 seconds to destroy the instance on the hypervisor. [ 972.087216] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 972.087216] env[65522]: DEBUG nova.compute.manager [-] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 972.087216] env[65522]: DEBUG nova.network.neutron [-] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 972.087216] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 972.087629] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 972.087889] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 972.154730] env[65522]: DEBUG nova.scheduler.client.report [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 972.159079] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 972.282337] env[65522]: DEBUG nova.compute.manager [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 972.282589] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 972.283570] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d3e9c5-eb37-4071-9d83-fa4e018628c0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.296113] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 972.296503] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7bd5be29-716c-4f14-8e45-216253dabc14 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.306268] env[65522]: DEBUG oslo_vmware.api [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 972.306268] env[65522]: value = "task-5114462" [ 972.306268] env[65522]: _type = "Task" [ 972.306268] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.316291] env[65522]: DEBUG oslo_vmware.api [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114462, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.430252] env[65522]: DEBUG oslo_vmware.api [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114461, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.35118} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.430559] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.430778] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 972.430959] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 972.431146] env[65522]: INFO nova.compute.manager [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Took 0.61 seconds to destroy the instance on the hypervisor. [ 972.431401] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 972.431664] env[65522]: DEBUG nova.compute.manager [-] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 972.431773] env[65522]: DEBUG nova.network.neutron [-] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 972.432039] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 972.432614] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 972.432880] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 972.450116] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82df9a65-c23f-44f6-a0b2-26f3e61d2b43 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "82a7ad85-a061-4eec-bd3f-fc977532dfbf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.819s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 972.550048] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 972.634881] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "82a7ad85-a061-4eec-bd3f-fc977532dfbf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 972.635038] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "82a7ad85-a061-4eec-bd3f-fc977532dfbf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 972.635389] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "82a7ad85-a061-4eec-bd3f-fc977532dfbf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 972.635472] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "82a7ad85-a061-4eec-bd3f-fc977532dfbf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 972.635624] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "82a7ad85-a061-4eec-bd3f-fc977532dfbf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 972.638390] env[65522]: INFO nova.compute.manager [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Terminating instance [ 972.663544] env[65522]: DEBUG oslo_concurrency.lockutils [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.433s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 972.664915] env[65522]: DEBUG nova.compute.manager [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 972.666881] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.520s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 972.670070] env[65522]: INFO nova.compute.claims [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 972.820155] env[65522]: DEBUG oslo_vmware.api [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114462, 'name': PowerOffVM_Task, 'duration_secs': 0.435357} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.820533] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 972.820834] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 972.820915] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e1c435bb-4ca5-44a8-ad60-c7fb6bdb6c32 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.896202] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 972.897643] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 972.898042] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleting the datastore file [datastore1] bb20d99f-fc49-47bf-a6b7-74ea966b71eb {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.898893] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0cccbeaa-ceb3-4681-b2d3-323297df3515 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.910850] env[65522]: DEBUG oslo_vmware.api [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for the task: (returnval){ [ 972.910850] env[65522]: value = "task-5114464" [ 972.910850] env[65522]: _type = "Task" [ 972.910850] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.925426] env[65522]: DEBUG oslo_vmware.api [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114464, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.971066] env[65522]: DEBUG nova.compute.manager [req-0b9d4df7-da90-4bac-9349-2683e0cd742e req-2ee7adf2-775c-4312-b6ae-6c56048e3b22 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Received event network-vif-deleted-3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 972.971293] env[65522]: INFO nova.compute.manager [req-0b9d4df7-da90-4bac-9349-2683e0cd742e req-2ee7adf2-775c-4312-b6ae-6c56048e3b22 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Neutron deleted interface 3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6; detaching it from the instance and deleting it from the info cache [ 972.971457] env[65522]: DEBUG nova.network.neutron [req-0b9d4df7-da90-4bac-9349-2683e0cd742e req-2ee7adf2-775c-4312-b6ae-6c56048e3b22 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 973.143458] env[65522]: DEBUG nova.compute.manager [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 973.143566] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 973.144750] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc172a0c-3786-4c9c-845b-090af96127e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.155976] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 973.156244] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d6473998-1280-4e29-a8df-21089fefb376 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.165279] env[65522]: DEBUG oslo_vmware.api [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 973.165279] env[65522]: value = "task-5114465" [ 973.165279] env[65522]: _type = "Task" [ 973.165279] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.176926] env[65522]: DEBUG nova.compute.utils [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 973.186385] env[65522]: DEBUG nova.compute.manager [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Not allocating networking since 'none' was specified. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 973.186876] env[65522]: DEBUG oslo_vmware.api [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114465, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.226828] env[65522]: DEBUG nova.network.neutron [-] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 973.424065] env[65522]: DEBUG oslo_vmware.api [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Task: {'id': task-5114464, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.322508} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.424480] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.424561] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 973.425549] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 973.425549] env[65522]: INFO nova.compute.manager [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Took 1.14 seconds to destroy the instance on the hypervisor. [ 973.425549] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 973.425549] env[65522]: DEBUG nova.compute.manager [-] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 973.425549] env[65522]: DEBUG nova.network.neutron [-] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 973.425869] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 973.427025] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 973.427025] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 973.478337] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-babec5e9-dc56-48c1-830a-aaebf922fe1a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.482371] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 973.497662] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ecfa1d5-bd02-42a8-a9c9-8ef35f87d4d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.533083] env[65522]: DEBUG nova.compute.manager [req-0b9d4df7-da90-4bac-9349-2683e0cd742e req-2ee7adf2-775c-4312-b6ae-6c56048e3b22 service nova] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Detach interface failed, port_id=3bbf4b9c-9f1c-49eb-a2e5-ce75414b91d6, reason: Instance 68d8d59b-4fc8-4993-8c9e-31b2329c88dd could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 973.569087] env[65522]: DEBUG nova.network.neutron [-] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 973.676881] env[65522]: DEBUG oslo_vmware.api [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114465, 'name': PowerOffVM_Task, 'duration_secs': 0.221957} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.676977] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 973.677216] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 973.677382] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e363a8aa-a9fd-4522-87a1-8cfa8475687f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.690634] env[65522]: DEBUG nova.compute.manager [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 973.738115] env[65522]: INFO nova.compute.manager [-] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Took 1.65 seconds to deallocate network for instance. [ 973.745708] env[65522]: DEBUG nova.compute.manager [req-884a681d-78db-4caa-b5ff-87f1ef2c4ed3 req-8219a8f5-d67b-4878-92fa-9fd74efb12ef service nova] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Received event network-vif-deleted-652babe2-668c-4364-92c2-8ceaba23c0ac {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 973.745708] env[65522]: INFO nova.compute.manager [req-884a681d-78db-4caa-b5ff-87f1ef2c4ed3 req-8219a8f5-d67b-4878-92fa-9fd74efb12ef service nova] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Neutron deleted interface 652babe2-668c-4364-92c2-8ceaba23c0ac; detaching it from the instance and deleting it from the info cache [ 973.745708] env[65522]: DEBUG nova.network.neutron [req-884a681d-78db-4caa-b5ff-87f1ef2c4ed3 req-8219a8f5-d67b-4878-92fa-9fd74efb12ef service nova] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 973.761678] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 973.765482] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 973.765482] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleting the datastore file [datastore2] 82a7ad85-a061-4eec-bd3f-fc977532dfbf {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 973.765482] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8789a94f-17cc-4969-b58d-951216fc03f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.780035] env[65522]: DEBUG oslo_vmware.api [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 973.780035] env[65522]: value = "task-5114467" [ 973.780035] env[65522]: _type = "Task" [ 973.780035] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.793375] env[65522]: DEBUG oslo_vmware.api [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114467, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.066023] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0010456a-efea-4776-8f66-8fc7b7c708ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.077122] env[65522]: INFO nova.compute.manager [-] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Took 1.65 seconds to deallocate network for instance. [ 974.080048] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f323431-e9c7-49fc-865a-84b670a0121c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.121912] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a7623f0-bc17-46a5-a859-c41349269791 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.130944] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b710acf7-ebf8-4289-a129-8a77663e8127 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.148100] env[65522]: DEBUG nova.compute.provider_tree [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.218831] env[65522]: DEBUG nova.network.neutron [-] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 974.251753] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 974.251991] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1bed8217-82d8-45db-bbb6-66f380ad8612 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.266282] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248b8ce7-72fc-4bc5-a79b-4783a7599dc9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.291286] env[65522]: DEBUG oslo_vmware.api [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114467, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.272967} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.291548] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 974.291726] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 974.291894] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 974.292074] env[65522]: INFO nova.compute.manager [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Took 1.15 seconds to destroy the instance on the hypervisor. [ 974.292321] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 974.292535] env[65522]: DEBUG nova.compute.manager [-] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 974.292612] env[65522]: DEBUG nova.network.neutron [-] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 974.293622] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 974.293622] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 974.293970] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 974.320565] env[65522]: DEBUG nova.compute.manager [req-884a681d-78db-4caa-b5ff-87f1ef2c4ed3 req-8219a8f5-d67b-4878-92fa-9fd74efb12ef service nova] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Detach interface failed, port_id=652babe2-668c-4364-92c2-8ceaba23c0ac, reason: Instance bb20d99f-fc49-47bf-a6b7-74ea966b71eb could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 974.355066] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 974.591134] env[65522]: DEBUG oslo_concurrency.lockutils [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 974.670021] env[65522]: ERROR nova.scheduler.client.report [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [req-f0a8aed6-b893-4f20-9ec9-768aa0a222b4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 9a0d1ac8-ed26-4657-8899-fbceab29db0d. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f0a8aed6-b893-4f20-9ec9-768aa0a222b4"}]} [ 974.687680] env[65522]: DEBUG nova.scheduler.client.report [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Refreshing inventories for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 974.701964] env[65522]: DEBUG nova.compute.manager [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 974.705512] env[65522]: DEBUG nova.scheduler.client.report [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Updating ProviderTree inventory for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 974.705747] env[65522]: DEBUG nova.compute.provider_tree [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.717467] env[65522]: DEBUG nova.scheduler.client.report [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Refreshing aggregate associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, aggregates: None {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 974.719813] env[65522]: INFO nova.compute.manager [-] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Took 1.29 seconds to deallocate network for instance. [ 974.731328] env[65522]: DEBUG nova.virt.hardware [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 974.731532] env[65522]: DEBUG nova.virt.hardware [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 974.731696] env[65522]: DEBUG nova.virt.hardware [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 974.731984] env[65522]: DEBUG nova.virt.hardware [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 974.732164] env[65522]: DEBUG nova.virt.hardware [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 974.732313] env[65522]: DEBUG nova.virt.hardware [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 974.732545] env[65522]: DEBUG nova.virt.hardware [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.732725] env[65522]: DEBUG nova.virt.hardware [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 974.732911] env[65522]: DEBUG nova.virt.hardware [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 974.733106] env[65522]: DEBUG nova.virt.hardware [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 974.733330] env[65522]: DEBUG nova.virt.hardware [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 974.734188] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95851573-d9d4-4fe1-ba1c-10882d87e6e4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.743564] env[65522]: DEBUG nova.scheduler.client.report [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Refreshing trait associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 974.746966] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c9f74fc-7d2f-475f-b9c2-bd4a224bf279 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.763685] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 974.769967] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Creating folder: Project (065915a22e8b4ae09628cc71c7a188a0). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 974.770102] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f7ed516f-fe12-4e85-85ca-03f1348f76f4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.783930] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Created folder: Project (065915a22e8b4ae09628cc71c7a188a0) in parent group-v994660. [ 974.784284] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Creating folder: Instances. Parent ref: group-v994871. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 974.784578] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-952e58bf-2ac0-424b-a4c4-c7433538d0d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.802197] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Created folder: Instances in parent group-v994871. [ 974.802464] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 974.802678] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 974.802926] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca3b7b26-a0d4-4326-934c-fcbfc011c89e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.825782] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 974.825782] env[65522]: value = "task-5114470" [ 974.825782] env[65522]: _type = "Task" [ 974.825782] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.835604] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114470, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.896545] env[65522]: INFO nova.compute.manager [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Rebuilding instance [ 974.944081] env[65522]: DEBUG nova.compute.manager [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 974.944985] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e74d1f2-b3d7-41ec-8fd4-3cce3f66dcaa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.014126] env[65522]: DEBUG nova.compute.manager [req-361a7466-a995-47c0-a9b3-9e89ad91ffec req-fa25f72d-67c2-4916-a893-7e80c7b6936d service nova] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Received event network-vif-deleted-dd06b43a-4102-4a7a-a82e-cf76b354ba13 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 975.108926] env[65522]: DEBUG nova.network.neutron [-] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 975.136801] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e95ec8-6d64-4a97-8065-4c5decec19c8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.148424] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e775999-bf50-44f1-a171-f480bf3fb761 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.186368] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b27834-cc03-4324-b024-d084d4808a88 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.198134] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98724a60-69e3-4023-b89e-f6a41dd3d844 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.215387] env[65522]: DEBUG nova.compute.provider_tree [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 975.239232] env[65522]: DEBUG oslo_concurrency.lockutils [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 975.338348] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114470, 'name': CreateVM_Task, 'duration_secs': 0.471919} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.338824] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 975.339409] env[65522]: DEBUG oslo_concurrency.lockutils [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.340031] env[65522]: DEBUG oslo_concurrency.lockutils [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 975.340031] env[65522]: DEBUG oslo_concurrency.lockutils [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 975.340323] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d33eee74-a279-4d63-870d-dcd0a12aea16 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.346802] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 975.346802] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f1b5ba-0501-a936-40d2-55da39bdc0cf" [ 975.346802] env[65522]: _type = "Task" [ 975.346802] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.357532] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f1b5ba-0501-a936-40d2-55da39bdc0cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.615169] env[65522]: INFO nova.compute.manager [-] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Took 1.32 seconds to deallocate network for instance. [ 975.753324] env[65522]: DEBUG nova.scheduler.client.report [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Updated inventory for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with generation 103 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 975.753625] env[65522]: DEBUG nova.compute.provider_tree [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Updating resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d generation from 103 to 104 during operation: update_inventory {{(pid=65522) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 975.753901] env[65522]: DEBUG nova.compute.provider_tree [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 975.860477] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f1b5ba-0501-a936-40d2-55da39bdc0cf, 'name': SearchDatastore_Task, 'duration_secs': 0.014282} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.860816] env[65522]: DEBUG oslo_concurrency.lockutils [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 975.861083] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 975.861326] env[65522]: DEBUG oslo_concurrency.lockutils [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.861471] env[65522]: DEBUG oslo_concurrency.lockutils [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 975.861686] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 975.861973] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0eba8220-f6ad-4a9b-9bf3-e7584a2d3604 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.871826] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 975.872017] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 975.872782] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03126300-c7f1-48c0-968c-a9c22987e2a0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.879299] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 975.879299] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5249ce9f-69d9-bf28-3bb1-11be53359aeb" [ 975.879299] env[65522]: _type = "Task" [ 975.879299] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.888860] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5249ce9f-69d9-bf28-3bb1-11be53359aeb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.963505] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 975.963854] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c7694908-891b-4083-9e2f-aca3bd617bfb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.972497] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 975.972497] env[65522]: value = "task-5114471" [ 975.972497] env[65522]: _type = "Task" [ 975.972497] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.984876] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114471, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.987247] env[65522]: DEBUG nova.compute.manager [req-f18b13f0-0a0d-44cd-b4c9-305be61c0ad2 req-5a65b8ae-f483-4246-a2ec-d4ed1364a853 service nova] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Received event network-vif-deleted-37eeac6b-6857-4309-b861-5eb05cdba980 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 976.160643] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 976.264287] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.597s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 976.265050] env[65522]: DEBUG nova.compute.manager [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 976.269147] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.936s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 976.269540] env[65522]: DEBUG nova.objects.instance [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lazy-loading 'resources' on Instance uuid 1eebb320-30e2-4d8b-b1fd-19e31fb35b77 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.392042] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5249ce9f-69d9-bf28-3bb1-11be53359aeb, 'name': SearchDatastore_Task, 'duration_secs': 0.014558} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.392683] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1310fff-9757-449d-ad7b-57686318e634 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.399202] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 976.399202] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5231833c-a7f6-187c-ad5d-eb34d7bfecf0" [ 976.399202] env[65522]: _type = "Task" [ 976.399202] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.408768] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5231833c-a7f6-187c-ad5d-eb34d7bfecf0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.483464] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114471, 'name': PowerOffVM_Task, 'duration_secs': 0.249135} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.485940] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 976.485940] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 976.485940] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2fad4e-47a7-4dd0-bc51-444a496e08ed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.493940] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 976.495171] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45408659-2ab9-4632-ae65-fa4da3613500 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.564310] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 976.564521] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 976.564739] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleting the datastore file [datastore1] 9c39b722-56ff-44fa-8f66-3e3432645a68 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 976.565082] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b5183c96-9831-4a08-862f-4c961b6f589e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.574128] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 976.574128] env[65522]: value = "task-5114473" [ 976.574128] env[65522]: _type = "Task" [ 976.574128] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.583917] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114473, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.779834] env[65522]: DEBUG nova.compute.utils [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 976.784329] env[65522]: DEBUG nova.compute.manager [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Not allocating networking since 'none' was specified. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 976.923604] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5231833c-a7f6-187c-ad5d-eb34d7bfecf0, 'name': SearchDatastore_Task, 'duration_secs': 0.015228} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.927428] env[65522]: DEBUG oslo_concurrency.lockutils [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 976.927879] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 95a10a6f-8c6c-435d-a360-dff10eee0855/95a10a6f-8c6c-435d-a360-dff10eee0855.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 976.930096] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-96305f33-c265-4599-86be-0798324c5b20 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.938947] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 976.938947] env[65522]: value = "task-5114474" [ 976.938947] env[65522]: _type = "Task" [ 976.938947] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.951803] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114474, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.086541] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114473, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.469049} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.089339] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 977.089540] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 977.089717] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 977.156818] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b211f75-218f-4723-99c9-d331f7d2fc27 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.167298] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627f25cf-76ae-435d-a4e3-1acf87c77fda {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.204157] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942467c3-c181-45ba-a21a-f5c7000f5536 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.215248] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea8f8d5-fe6b-4b95-9f24-e97df78849fd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.234188] env[65522]: DEBUG nova.compute.provider_tree [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.249963] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52798a94-6485-7710-c375-de9f93b9b1ea/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 977.251200] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49035d0a-b027-4703-8f53-c4386d031368 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.261726] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52798a94-6485-7710-c375-de9f93b9b1ea/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 977.261726] env[65522]: ERROR oslo_vmware.rw_handles [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52798a94-6485-7710-c375-de9f93b9b1ea/disk-0.vmdk due to incomplete transfer. [ 977.261726] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1f0968d1-27fc-4849-8d06-c694fe84371e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.270234] env[65522]: DEBUG oslo_vmware.rw_handles [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52798a94-6485-7710-c375-de9f93b9b1ea/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 977.270438] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Uploaded image 57ba37b7-4542-434d-8a76-5c4f6d75462e to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 977.273892] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 977.274150] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9b563d4b-ee72-4503-bf19-78b787a00f7b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.282958] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 977.282958] env[65522]: value = "task-5114475" [ 977.282958] env[65522]: _type = "Task" [ 977.282958] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.287043] env[65522]: DEBUG nova.compute.manager [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 977.297167] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114475, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.456255] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114474, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.737685] env[65522]: DEBUG nova.scheduler.client.report [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 977.800601] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114475, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.958044] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114474, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.740462} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.958044] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 95a10a6f-8c6c-435d-a360-dff10eee0855/95a10a6f-8c6c-435d-a360-dff10eee0855.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 977.958044] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 977.958044] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-909db335-d086-4892-8403-3a4bde5205af {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.966128] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 977.966128] env[65522]: value = "task-5114476" [ 977.966128] env[65522]: _type = "Task" [ 977.966128] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.981986] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114476, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.134366] env[65522]: DEBUG nova.virt.hardware [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 978.134635] env[65522]: DEBUG nova.virt.hardware [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 978.134791] env[65522]: DEBUG nova.virt.hardware [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 978.134967] env[65522]: DEBUG nova.virt.hardware [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 978.135284] env[65522]: DEBUG nova.virt.hardware [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 978.135531] env[65522]: DEBUG nova.virt.hardware [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 978.135763] env[65522]: DEBUG nova.virt.hardware [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 978.136019] env[65522]: DEBUG nova.virt.hardware [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 978.136240] env[65522]: DEBUG nova.virt.hardware [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 978.136494] env[65522]: DEBUG nova.virt.hardware [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 978.136650] env[65522]: DEBUG nova.virt.hardware [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 978.137674] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738c0ccb-38a3-4809-9620-e08c57b27886 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.147303] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbcaff4b-61b3-4660-ae72-ba9b03e63a8f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.165039] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:7b:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4307c18-b235-43cd-bcd5-e226012d8ee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '54206d03-4403-4bca-8ba8-1f4e88682cb6', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 978.172688] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 978.173108] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 978.173383] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-13f1faa1-9942-4adf-8947-e7e9b07fc56f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.197893] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 978.197893] env[65522]: value = "task-5114477" [ 978.197893] env[65522]: _type = "Task" [ 978.197893] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.210079] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114477, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.219686] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "8da2a500-6f0e-4eda-9a92-79510753b3b8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.219958] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "8da2a500-6f0e-4eda-9a92-79510753b3b8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.220180] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "8da2a500-6f0e-4eda-9a92-79510753b3b8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.220356] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "8da2a500-6f0e-4eda-9a92-79510753b3b8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.220526] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "8da2a500-6f0e-4eda-9a92-79510753b3b8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.223415] env[65522]: INFO nova.compute.manager [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Terminating instance [ 978.247967] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.976s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.247967] env[65522]: DEBUG oslo_concurrency.lockutils [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.518s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.247967] env[65522]: DEBUG nova.objects.instance [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'resources' on Instance uuid 9df1a51a-2811-4486-a4c6-58d618f2ae7d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.276878] env[65522]: INFO nova.scheduler.client.report [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Deleted allocations for instance 1eebb320-30e2-4d8b-b1fd-19e31fb35b77 [ 978.296591] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114475, 'name': Destroy_Task, 'duration_secs': 0.748741} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.296961] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Destroyed the VM [ 978.297239] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 978.297546] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9f18a51d-6d03-425e-90b8-2a9d5ff7b1db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.302557] env[65522]: DEBUG nova.compute.manager [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 978.306612] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 978.306612] env[65522]: value = "task-5114478" [ 978.306612] env[65522]: _type = "Task" [ 978.306612] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.320041] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114478, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.340536] env[65522]: DEBUG nova.virt.hardware [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 978.340900] env[65522]: DEBUG nova.virt.hardware [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 978.341236] env[65522]: DEBUG nova.virt.hardware [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 978.341462] env[65522]: DEBUG nova.virt.hardware [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 978.341685] env[65522]: DEBUG nova.virt.hardware [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 978.341867] env[65522]: DEBUG nova.virt.hardware [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 978.342099] env[65522]: DEBUG nova.virt.hardware [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 978.342263] env[65522]: DEBUG nova.virt.hardware [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 978.342480] env[65522]: DEBUG nova.virt.hardware [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 978.342605] env[65522]: DEBUG nova.virt.hardware [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 978.342778] env[65522]: DEBUG nova.virt.hardware [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 978.344077] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a1b6494-9950-416a-bf6a-043cd84ef8e1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.353201] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcdf4b0b-924b-4356-822c-b972fa328eff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.370373] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 978.382736] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 978.382736] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 978.382736] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a6fb0927-6e4c-4d39-91df-50894b7e73e3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.400296] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 978.400296] env[65522]: value = "task-5114479" [ 978.400296] env[65522]: _type = "Task" [ 978.400296] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.410465] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114479, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.479550] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114476, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07844} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.480050] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 978.481069] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b7aac8-8636-4379-8bbd-5afa10d8f99f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.506198] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 95a10a6f-8c6c-435d-a360-dff10eee0855/95a10a6f-8c6c-435d-a360-dff10eee0855.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.506802] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9b0fef1-9c78-49e1-b892-e0ba239076fe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.531726] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 978.531726] env[65522]: value = "task-5114480" [ 978.531726] env[65522]: _type = "Task" [ 978.531726] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.542510] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114480, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.629979] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 978.630359] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 978.717753] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114477, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.733061] env[65522]: DEBUG nova.compute.manager [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 978.733380] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 978.734409] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d23f55-75a5-48df-a0b6-8aea57af8fec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.743666] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 978.744051] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ae6d522-8201-4706-b70f-4cc4a3c9a389 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.793332] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82977a89-1661-47ed-8a70-2a7df5987123 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "1eebb320-30e2-4d8b-b1fd-19e31fb35b77" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.679s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.821692] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114478, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.917807] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114479, 'name': CreateVM_Task, 'duration_secs': 0.507352} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.918431] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 978.919435] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.919435] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 978.920460] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 978.921919] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc8da5ad-22c9-4324-8ae3-d44f66f6cd79 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.929471] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 978.929471] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a44e2e-71fd-4d66-9289-295a2362b590" [ 978.929471] env[65522]: _type = "Task" [ 978.929471] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.940706] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a44e2e-71fd-4d66-9289-295a2362b590, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.051132] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114480, 'name': ReconfigVM_Task, 'duration_secs': 0.46082} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.051473] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 95a10a6f-8c6c-435d-a360-dff10eee0855/95a10a6f-8c6c-435d-a360-dff10eee0855.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.052127] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-732b0099-bc3a-428b-ae33-479dbb6c80b4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.063481] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 979.063481] env[65522]: value = "task-5114482" [ 979.063481] env[65522]: _type = "Task" [ 979.063481] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.078976] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114482, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.092588] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "8859b051-8f75-4aad-b789-42662019d4c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 979.092969] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "8859b051-8f75-4aad-b789-42662019d4c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 979.151220] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 979.151277] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 979.151492] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 979.151768] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 979.153684] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 979.154207] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 979.154316] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 979.154502] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 979.206328] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d02f445-d13f-4b9f-ad3a-4ba96fb4bb96 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.227374] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28411ced-94f5-41d0-a1e4-2a252a1985ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.231885] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114477, 'name': CreateVM_Task, 'duration_secs': 0.618625} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.232396] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 979.233372] env[65522]: WARNING neutronclient.v2_0.client [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 979.233849] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.269263] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d3cbb60-d109-48de-aeb9-9a94d8b9538c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.272502] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 979.272751] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 979.272909] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleting the datastore file [datastore1] 8da2a500-6f0e-4eda-9a92-79510753b3b8 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.273216] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-151702c2-f822-42d7-9043-0403bea90a67 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.283926] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a746d5-033d-4766-9808-54b77834c5a1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.288332] env[65522]: DEBUG oslo_vmware.api [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 979.288332] env[65522]: value = "task-5114483" [ 979.288332] env[65522]: _type = "Task" [ 979.288332] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.302025] env[65522]: DEBUG nova.compute.provider_tree [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.307568] env[65522]: DEBUG oslo_vmware.api [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114483, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.319016] env[65522]: DEBUG oslo_vmware.api [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114478, 'name': RemoveSnapshot_Task, 'duration_secs': 0.791789} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.319295] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 979.319512] env[65522]: INFO nova.compute.manager [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Took 18.70 seconds to snapshot the instance on the hypervisor. [ 979.441752] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a44e2e-71fd-4d66-9289-295a2362b590, 'name': SearchDatastore_Task, 'duration_secs': 0.014202} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.442304] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 979.442724] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.443142] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.443426] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 979.443707] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.444136] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 979.444567] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 979.444953] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ba024e3-feb9-4c6e-b100-5cd2bfa455ed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.447725] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4acb8c7-c14d-4ccf-ad5c-a7c4122ef7fb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.457239] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 979.457239] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dc9014-4b8c-3866-8ee8-a58715c28c24" [ 979.457239] env[65522]: _type = "Task" [ 979.457239] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.459421] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.459856] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 979.461269] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ccc8729-15ee-420e-a019-c751e74cfe8a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.469566] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dc9014-4b8c-3866-8ee8-a58715c28c24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.477018] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 979.477018] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5211bf0b-ba34-11f3-98c6-dd17065a2c5b" [ 979.477018] env[65522]: _type = "Task" [ 979.477018] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.488055] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5211bf0b-ba34-11f3-98c6-dd17065a2c5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.574467] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114482, 'name': Rename_Task, 'duration_secs': 0.191175} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.574677] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 979.575207] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22cc7fda-1e7f-49fd-8fa8-f812d67331a7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.583774] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 979.583774] env[65522]: value = "task-5114484" [ 979.583774] env[65522]: _type = "Task" [ 979.583774] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.593882] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114484, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.598890] env[65522]: DEBUG nova.compute.manager [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 979.601811] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 979.602563] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 979.658501] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 979.803352] env[65522]: DEBUG oslo_vmware.api [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114483, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187875} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.804669] env[65522]: DEBUG nova.scheduler.client.report [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 979.810131] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.810357] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 979.810592] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 979.810907] env[65522]: INFO nova.compute.manager [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Took 1.08 seconds to destroy the instance on the hypervisor. [ 979.811079] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 979.811327] env[65522]: DEBUG nova.compute.manager [-] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 979.811427] env[65522]: DEBUG nova.network.neutron [-] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 979.811789] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 979.812990] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 979.813275] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 979.824048] env[65522]: DEBUG nova.compute.manager [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Instance disappeared during snapshot {{(pid=65522) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4652}} [ 979.845113] env[65522]: DEBUG nova.compute.manager [None req-d2b10747-b24f-4ef0-a302-ac14a926523a tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Image not found during clean up 57ba37b7-4542-434d-8a76-5c4f6d75462e {{(pid=65522) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4658}} [ 979.859837] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 979.966014] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dc9014-4b8c-3866-8ee8-a58715c28c24, 'name': SearchDatastore_Task, 'duration_secs': 0.016738} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.966340] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 979.966569] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.966780] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.986483] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5211bf0b-ba34-11f3-98c6-dd17065a2c5b, 'name': SearchDatastore_Task, 'duration_secs': 0.015669} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.986913] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69594f7e-a4dd-4a3f-8224-f9471d6b01c5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.993541] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 979.993541] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cae645-57d9-d112-011d-e26ac040fb1e" [ 979.993541] env[65522]: _type = "Task" [ 979.993541] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.003351] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cae645-57d9-d112-011d-e26ac040fb1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.095487] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114484, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.112387] env[65522]: DEBUG nova.compute.manager [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 980.141464] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 980.163406] env[65522]: DEBUG nova.compute.manager [req-87ffc615-18a9-4591-ac6f-58260d429df7 req-3d94ba07-882e-4cbd-a9f3-6da4fb568986 service nova] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Received event network-vif-deleted-5c7d9e50-0f3f-477e-af8d-86a5fe295457 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 980.164574] env[65522]: INFO nova.compute.manager [req-87ffc615-18a9-4591-ac6f-58260d429df7 req-3d94ba07-882e-4cbd-a9f3-6da4fb568986 service nova] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Neutron deleted interface 5c7d9e50-0f3f-477e-af8d-86a5fe295457; detaching it from the instance and deleting it from the info cache [ 980.164574] env[65522]: DEBUG nova.network.neutron [req-87ffc615-18a9-4591-ac6f-58260d429df7 req-3d94ba07-882e-4cbd-a9f3-6da4fb568986 service nova] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 980.326356] env[65522]: DEBUG oslo_concurrency.lockutils [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.078s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 980.328664] env[65522]: DEBUG oslo_concurrency.lockutils [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 10.494s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 980.350915] env[65522]: INFO nova.scheduler.client.report [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Deleted allocations for instance 9df1a51a-2811-4486-a4c6-58d618f2ae7d [ 980.506498] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cae645-57d9-d112-011d-e26ac040fb1e, 'name': SearchDatastore_Task, 'duration_secs': 0.023672} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.506957] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 980.507069] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 7701b92d-805e-4837-b900-326910b5eef5/7701b92d-805e-4837-b900-326910b5eef5.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 980.507411] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 980.507491] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 980.507715] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49f323de-1140-49db-b443-e2706515c5d9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.510030] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d4b71d60-01aa-41fa-8d46-8bacd756ed12 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.518057] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 980.518057] env[65522]: value = "task-5114486" [ 980.518057] env[65522]: _type = "Task" [ 980.518057] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.522728] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 980.522929] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 980.524107] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5818b383-29cc-4aaa-8dc3-09603b916038 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.530615] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114486, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.534646] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 980.534646] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521ea22c-d1e4-0271-c5c7-eb1be3ae1a02" [ 980.534646] env[65522]: _type = "Task" [ 980.534646] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.543656] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521ea22c-d1e4-0271-c5c7-eb1be3ae1a02, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.577112] env[65522]: DEBUG nova.network.neutron [-] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 980.596253] env[65522]: DEBUG oslo_vmware.api [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114484, 'name': PowerOnVM_Task, 'duration_secs': 0.545792} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.596546] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 980.596745] env[65522]: INFO nova.compute.manager [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Took 5.89 seconds to spawn the instance on the hypervisor. [ 980.596924] env[65522]: DEBUG nova.compute.manager [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 980.597802] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d7826a-599f-44a8-b4a8-3f61dfeee4be {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.643405] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 980.668296] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-afc707d2-dc8c-4f4b-bc62-a87e17f37de6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.684012] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430d87e6-c606-462a-97d5-366748267866 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.716017] env[65522]: DEBUG nova.compute.manager [req-87ffc615-18a9-4591-ac6f-58260d429df7 req-3d94ba07-882e-4cbd-a9f3-6da4fb568986 service nova] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Detach interface failed, port_id=5c7d9e50-0f3f-477e-af8d-86a5fe295457, reason: Instance 8da2a500-6f0e-4eda-9a92-79510753b3b8 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 980.859530] env[65522]: DEBUG oslo_concurrency.lockutils [None req-598525c8-c444-4385-8329-bbfd0e22d618 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "9df1a51a-2811-4486-a4c6-58d618f2ae7d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.777s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 981.029141] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114486, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.049927] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521ea22c-d1e4-0271-c5c7-eb1be3ae1a02, 'name': SearchDatastore_Task, 'duration_secs': 0.025571} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.050694] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acf0a75d-4add-4084-86a9-4e0bd2374746 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.057089] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 981.057089] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52077511-8ce4-733d-5d32-a73f0111b40b" [ 981.057089] env[65522]: _type = "Task" [ 981.057089] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.071494] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52077511-8ce4-733d-5d32-a73f0111b40b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.079375] env[65522]: INFO nova.compute.manager [-] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Took 1.27 seconds to deallocate network for instance. [ 981.122357] env[65522]: INFO nova.compute.manager [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Took 18.63 seconds to build instance. [ 981.187718] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5483276-d01e-45f1-9807-7380805ec6d8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.198587] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e99ac0-df8b-4e87-8ccd-7c5227e264f5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.244590] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f74e94-d5e7-4cc5-a837-91d81a2a40e9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.255681] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940fd223-27fa-4f75-b624-f476db3bdd0d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.275066] env[65522]: DEBUG nova.compute.provider_tree [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.530290] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114486, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.778412] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52077511-8ce4-733d-5d32-a73f0111b40b, 'name': SearchDatastore_Task, 'duration_secs': 0.019643} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.778412] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 981.778412] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 9c39b722-56ff-44fa-8f66-3e3432645a68/9c39b722-56ff-44fa-8f66-3e3432645a68.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 981.778412] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b602bd0f-2490-4eb5-a397-643f4c0f2c65 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.778412] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 981.778412] env[65522]: value = "task-5114487" [ 981.778412] env[65522]: _type = "Task" [ 981.778412] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.778412] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114487, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.778412] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 981.778412] env[65522]: DEBUG oslo_concurrency.lockutils [None req-92f049ee-34be-41a5-b17a-f0d8a099e72c tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "95a10a6f-8c6c-435d-a360-dff10eee0855" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.135s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 981.779091] env[65522]: DEBUG nova.scheduler.client.report [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 982.030793] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114486, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.275143} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.032316] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 7701b92d-805e-4837-b900-326910b5eef5/7701b92d-805e-4837-b900-326910b5eef5.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 982.032592] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 982.032872] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f3b20d6-f916-4ec0-adb4-e6c5f14a824b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.046283] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 982.046283] env[65522]: value = "task-5114488" [ 982.046283] env[65522]: _type = "Task" [ 982.046283] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.062280] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114488, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.092020] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114487, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.558865] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114488, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084907} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.559190] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 982.560120] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b57f13ae-8d53-46e1-801b-735cf68cecae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.585698] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 7701b92d-805e-4837-b900-326910b5eef5/7701b92d-805e-4837-b900-326910b5eef5.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 982.586692] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4259f96d-af44-4d76-be75-76abc11fca59 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.610360] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114487, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.614266] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 982.614266] env[65522]: value = "task-5114489" [ 982.614266] env[65522]: _type = "Task" [ 982.614266] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.793507] env[65522]: DEBUG oslo_concurrency.lockutils [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.465s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 982.798078] env[65522]: DEBUG oslo_concurrency.lockutils [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 12.476s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 982.798295] env[65522]: DEBUG nova.objects.instance [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 983.098195] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114487, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.048851} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.099098] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 9c39b722-56ff-44fa-8f66-3e3432645a68/9c39b722-56ff-44fa-8f66-3e3432645a68.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 983.099098] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 983.099217] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-46a6b9cb-fef4-4880-ae7b-e3b92540a11f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.107720] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 983.107720] env[65522]: value = "task-5114490" [ 983.107720] env[65522]: _type = "Task" [ 983.107720] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.123770] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114490, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.130027] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114489, 'name': ReconfigVM_Task, 'duration_secs': 0.350854} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.130176] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 7701b92d-805e-4837-b900-326910b5eef5/7701b92d-805e-4837-b900-326910b5eef5.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 983.131879] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8376083a-305b-4fb7-8f60-da00800023de {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.138747] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 983.138747] env[65522]: value = "task-5114491" [ 983.138747] env[65522]: _type = "Task" [ 983.138747] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.148944] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114491, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.369234] env[65522]: INFO nova.scheduler.client.report [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleted allocation for migration 43f083bf-03be-46c8-ba8b-81c6563136b6 [ 983.544036] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "87a3c63d-794d-44ab-bad6-65c323d72ae7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.544451] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "87a3c63d-794d-44ab-bad6-65c323d72ae7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 983.544572] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "87a3c63d-794d-44ab-bad6-65c323d72ae7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.544682] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "87a3c63d-794d-44ab-bad6-65c323d72ae7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 983.544850] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "87a3c63d-794d-44ab-bad6-65c323d72ae7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 983.548125] env[65522]: INFO nova.compute.manager [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Terminating instance [ 983.620049] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114490, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076984} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.620428] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 983.621358] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2790da5a-9586-414b-928a-4eb82f3c2033 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.646253] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 9c39b722-56ff-44fa-8f66-3e3432645a68/9c39b722-56ff-44fa-8f66-3e3432645a68.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 983.646825] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbf2a50e-f552-4277-9205-d216f893462e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.672514] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114491, 'name': Rename_Task, 'duration_secs': 0.165036} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.676155] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 983.676155] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 983.676155] env[65522]: value = "task-5114492" [ 983.676155] env[65522]: _type = "Task" [ 983.676155] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.676155] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3b18eba-07bb-4c5b-b2cb-5c7699755a92 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.685538] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114492, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.686904] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 983.686904] env[65522]: value = "task-5114493" [ 983.686904] env[65522]: _type = "Task" [ 983.686904] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.697565] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114493, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.806531] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.808021] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 983.814921] env[65522]: DEBUG oslo_concurrency.lockutils [None req-144be6eb-84f2-4bd0-b7af-c01aaeb9643b tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 983.818084] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.247s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 983.818084] env[65522]: DEBUG nova.objects.instance [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lazy-loading 'resources' on Instance uuid 1b2779f5-c6e8-4226-a819-0560d63bd7df {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.877653] env[65522]: DEBUG oslo_concurrency.lockutils [None req-af93767e-3bdd-4c9c-ad7f-26137a583d89 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 18.819s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 984.053407] env[65522]: DEBUG nova.compute.manager [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 984.053697] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 984.054703] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c55038-24dd-4243-afac-0ba23c185a17 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.066693] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.067056] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9530964b-45ae-46c3-9d3b-0edb3e88cc05 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.075358] env[65522]: DEBUG oslo_vmware.api [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 984.075358] env[65522]: value = "task-5114494" [ 984.075358] env[65522]: _type = "Task" [ 984.075358] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.085483] env[65522]: DEBUG oslo_vmware.api [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114494, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.198554] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114492, 'name': ReconfigVM_Task, 'duration_secs': 0.434434} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.199463] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 9c39b722-56ff-44fa-8f66-3e3432645a68/9c39b722-56ff-44fa-8f66-3e3432645a68.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 984.200778] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d535dbd-a312-4399-bcbf-842aa5e6a3fe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.208677] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114493, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.214019] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 984.214019] env[65522]: value = "task-5114495" [ 984.214019] env[65522]: _type = "Task" [ 984.214019] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.226597] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114495, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.310977] env[65522]: DEBUG nova.compute.manager [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 984.353871] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "e7baf2a0-21dd-4610-8230-81b99b64856b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 984.354207] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "e7baf2a0-21dd-4610-8230-81b99b64856b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 984.588731] env[65522]: DEBUG oslo_vmware.api [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114494, 'name': PowerOffVM_Task, 'duration_secs': 0.301796} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.589057] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 984.589399] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 984.589706] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2ded24ee-20a3-4cff-8b1d-380042b4c4b3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.632273] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579b4a48-90bf-4c33-9818-2815886a8fdd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.641402] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8a14a8-61fa-4b76-a684-a0c40dbfe336 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.679711] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57c8a51-5597-4d5e-9a70-c151771a7f73 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.682949] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 984.683234] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 984.683454] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Deleting the datastore file [datastore2] 87a3c63d-794d-44ab-bad6-65c323d72ae7 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.683865] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 984.684133] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 984.684374] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 984.685033] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 984.685033] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 984.686775] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d15556e1-8e1a-437e-a432-4e5c5b17cc15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.690852] env[65522]: INFO nova.compute.manager [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Terminating instance [ 984.700692] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4886b2b-d65c-4ecc-a443-0300fa31ecb2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.706061] env[65522]: DEBUG oslo_vmware.api [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 984.706061] env[65522]: value = "task-5114497" [ 984.706061] env[65522]: _type = "Task" [ 984.706061] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.709066] env[65522]: DEBUG oslo_vmware.api [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114493, 'name': PowerOnVM_Task, 'duration_secs': 0.532198} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.713575] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 984.713831] env[65522]: INFO nova.compute.manager [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Took 6.41 seconds to spawn the instance on the hypervisor. [ 984.713989] env[65522]: DEBUG nova.compute.manager [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 984.723325] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a50a1049-9ba8-4dcc-add3-cd904f7f06b5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.726345] env[65522]: DEBUG nova.compute.provider_tree [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.736597] env[65522]: DEBUG oslo_vmware.api [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114497, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.746223] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114495, 'name': Rename_Task, 'duration_secs': 0.205181} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.746223] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 984.746346] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4c0ac8b9-1f3e-4875-9e04-012593a8fb60 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.753769] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 984.753769] env[65522]: value = "task-5114498" [ 984.753769] env[65522]: _type = "Task" [ 984.753769] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.763453] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114498, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.834357] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 984.863294] env[65522]: DEBUG nova.compute.manager [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 985.211733] env[65522]: DEBUG nova.compute.manager [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 985.211733] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 985.211733] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd2331e9-bb1a-4cb6-b573-f2e7d7815d7d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.229796] env[65522]: DEBUG oslo_vmware.api [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114497, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.229796] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 985.229939] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37668fbd-9bb0-4c7e-9b38-d4c5a072bbee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.236460] env[65522]: DEBUG nova.scheduler.client.report [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 985.249476] env[65522]: DEBUG oslo_vmware.api [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 985.249476] env[65522]: value = "task-5114499" [ 985.249476] env[65522]: _type = "Task" [ 985.249476] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.256076] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 985.256375] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 985.263093] env[65522]: INFO nova.compute.manager [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Took 22.14 seconds to build instance. [ 985.281737] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114498, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.285365] env[65522]: DEBUG oslo_vmware.api [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114499, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.388717] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 985.721071] env[65522]: DEBUG oslo_vmware.api [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114497, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.678801} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.721411] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.721516] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 985.721727] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 985.721919] env[65522]: INFO nova.compute.manager [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Took 1.67 seconds to destroy the instance on the hypervisor. [ 985.722182] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 985.722381] env[65522]: DEBUG nova.compute.manager [-] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 985.722478] env[65522]: DEBUG nova.network.neutron [-] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 985.722720] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 985.723285] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 985.723554] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 985.742202] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.925s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 985.745266] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.434s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 985.746814] env[65522]: INFO nova.compute.claims [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 985.767213] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 985.769583] env[65522]: DEBUG nova.compute.manager [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 985.773886] env[65522]: INFO nova.scheduler.client.report [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleted allocations for instance 1b2779f5-c6e8-4226-a819-0560d63bd7df [ 985.774924] env[65522]: INFO nova.compute.manager [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Rebuilding instance [ 985.777146] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2cd5d91-25b4-4ff8-a4da-4758a2f88e10 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "7701b92d-805e-4837-b900-326910b5eef5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.661s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 985.788723] env[65522]: DEBUG oslo_vmware.api [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114499, 'name': PowerOffVM_Task, 'duration_secs': 0.436382} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.789293] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 985.789497] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 985.789762] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-288b6e79-d244-451c-b64d-428c467bd11d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.795145] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114498, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.828151] env[65522]: DEBUG nova.compute.manager [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 985.829190] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b049aa2-ea42-4589-a721-31115dadb42b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.858036] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 985.858036] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 985.858228] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleting the datastore file [datastore1] a2a5fb2d-62e0-4809-a01f-f1df66dca58f {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 985.858421] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b465f9c9-7171-4d8e-b3bc-0f5c75bff340 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.865925] env[65522]: DEBUG oslo_vmware.api [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 985.865925] env[65522]: value = "task-5114501" [ 985.865925] env[65522]: _type = "Task" [ 985.865925] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.876046] env[65522]: DEBUG oslo_vmware.api [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114501, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.082914] env[65522]: DEBUG nova.compute.manager [req-6c58d6ac-4bde-4723-8350-268f2d945ef1 req-ed0854e9-90c8-473a-ad4e-7b06354a1daf service nova] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Received event network-vif-deleted-60d5d352-4fed-4b0f-9efc-ce13e2066444 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 986.082914] env[65522]: INFO nova.compute.manager [req-6c58d6ac-4bde-4723-8350-268f2d945ef1 req-ed0854e9-90c8-473a-ad4e-7b06354a1daf service nova] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Neutron deleted interface 60d5d352-4fed-4b0f-9efc-ce13e2066444; detaching it from the instance and deleting it from the info cache [ 986.082914] env[65522]: DEBUG nova.network.neutron [req-6c58d6ac-4bde-4723-8350-268f2d945ef1 req-ed0854e9-90c8-473a-ad4e-7b06354a1daf service nova] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 986.287548] env[65522]: DEBUG oslo_vmware.api [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114498, 'name': PowerOnVM_Task, 'duration_secs': 1.3275} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.287973] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 986.288026] env[65522]: DEBUG nova.compute.manager [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 986.289029] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8a4eb3-1853-4757-9844-59efe918a9cb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.296795] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a6c8da21-3dfa-41f7-ba79-0ad1507ed33b tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "1b2779f5-c6e8-4226-a819-0560d63bd7df" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.398s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 986.314520] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 986.377285] env[65522]: DEBUG oslo_vmware.api [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114501, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.382264} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.377543] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 986.377758] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 986.377950] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 986.378131] env[65522]: INFO nova.compute.manager [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Took 1.17 seconds to destroy the instance on the hypervisor. [ 986.378621] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 986.378621] env[65522]: DEBUG nova.compute.manager [-] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 986.378986] env[65522]: DEBUG nova.network.neutron [-] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 986.379110] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 986.379656] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 986.379944] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 986.429810] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 986.494711] env[65522]: DEBUG nova.network.neutron [-] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 986.585028] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-54d2fab8-da40-4ebf-807c-785beed03fa9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.602209] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb57c033-1c75-4890-b38d-f4c8615a0808 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.634142] env[65522]: DEBUG nova.compute.manager [req-6c58d6ac-4bde-4723-8350-268f2d945ef1 req-ed0854e9-90c8-473a-ad4e-7b06354a1daf service nova] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Detach interface failed, port_id=60d5d352-4fed-4b0f-9efc-ce13e2066444, reason: Instance 87a3c63d-794d-44ab-bad6-65c323d72ae7 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 986.661806] env[65522]: DEBUG nova.compute.manager [req-4d1e983a-764a-4ea3-8151-641bf9139b85 req-6be348b8-0425-4ffb-a24a-1a5b1a6cb9fe service nova] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Received event network-vif-deleted-f8145f30-4e64-4a22-89a7-ffa393891ffb {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 986.662060] env[65522]: INFO nova.compute.manager [req-4d1e983a-764a-4ea3-8151-641bf9139b85 req-6be348b8-0425-4ffb-a24a-1a5b1a6cb9fe service nova] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Neutron deleted interface f8145f30-4e64-4a22-89a7-ffa393891ffb; detaching it from the instance and deleting it from the info cache [ 986.662260] env[65522]: DEBUG nova.network.neutron [req-4d1e983a-764a-4ea3-8151-641bf9139b85 req-6be348b8-0425-4ffb-a24a-1a5b1a6cb9fe service nova] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 986.819450] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 986.845787] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 986.846414] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1896e065-9c44-48b6-b022-58abf5824f56 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.855609] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 986.855609] env[65522]: value = "task-5114502" [ 986.855609] env[65522]: _type = "Task" [ 986.855609] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.867450] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114502, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.999111] env[65522]: INFO nova.compute.manager [-] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Took 1.28 seconds to deallocate network for instance. [ 987.108760] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff542e0-68e8-4144-a365-5cdce5eec2e9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.119758] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd8c7fbd-d391-4944-a6f8-696df323a15f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.152576] env[65522]: DEBUG nova.network.neutron [-] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 987.155187] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c959174b-9577-4a38-af4c-8337c30f1e1c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.164777] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1a373c-f7cd-4de1-94fb-a62a38f6ce22 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.169588] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-173fb3de-407a-4c78-8227-906d0e49b8e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.183363] env[65522]: DEBUG nova.compute.provider_tree [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.188719] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff630df-4010-484a-bddc-2278b2be9e68 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.231773] env[65522]: DEBUG nova.compute.manager [req-4d1e983a-764a-4ea3-8151-641bf9139b85 req-6be348b8-0425-4ffb-a24a-1a5b1a6cb9fe service nova] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Detach interface failed, port_id=f8145f30-4e64-4a22-89a7-ffa393891ffb, reason: Instance a2a5fb2d-62e0-4809-a01f-f1df66dca58f could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 987.367237] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114502, 'name': PowerOffVM_Task, 'duration_secs': 0.287035} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.367237] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 987.367237] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 987.367866] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d3dfb9-6875-4c6b-88fb-afbde7d71da5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.377212] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 987.377212] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6f4d6768-d476-43fc-8da3-f79365968605 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.405153] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 987.405250] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 987.405765] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Deleting the datastore file [datastore2] 7701b92d-805e-4837-b900-326910b5eef5 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.406524] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eefbd0f5-6b68-4216-8f58-9794161097bf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.416845] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 987.416845] env[65522]: value = "task-5114504" [ 987.416845] env[65522]: _type = "Task" [ 987.416845] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.427549] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114504, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.509373] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 987.655495] env[65522]: INFO nova.compute.manager [-] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Took 1.28 seconds to deallocate network for instance. [ 987.687591] env[65522]: DEBUG nova.scheduler.client.report [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 987.928016] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114504, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.235407} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.928284] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 987.928467] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 987.928674] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 987.940310] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "43691901-7ac0-4523-a19c-eedf5ae4c040" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 987.940717] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "43691901-7ac0-4523-a19c-eedf5ae4c040" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 988.162708] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 988.193038] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 988.193446] env[65522]: DEBUG nova.compute.manager [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 988.196306] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.945s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 988.196604] env[65522]: DEBUG nova.objects.instance [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Lazy-loading 'resources' on Instance uuid 68d8d59b-4fc8-4993-8c9e-31b2329c88dd {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.445566] env[65522]: DEBUG nova.compute.manager [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 988.700134] env[65522]: DEBUG nova.compute.utils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 988.703885] env[65522]: DEBUG nova.compute.manager [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 988.704153] env[65522]: DEBUG nova.network.neutron [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 988.704604] env[65522]: WARNING neutronclient.v2_0.client [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 988.704765] env[65522]: WARNING neutronclient.v2_0.client [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 988.705355] env[65522]: WARNING openstack [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 988.705766] env[65522]: WARNING openstack [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 988.763338] env[65522]: DEBUG nova.policy [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8f9627246e914e999f9acd463f8162c9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '535709df1c2542f9a9d486d7153af6ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 988.972134] env[65522]: DEBUG nova.virt.hardware [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 988.972134] env[65522]: DEBUG nova.virt.hardware [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 988.972134] env[65522]: DEBUG nova.virt.hardware [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 988.972134] env[65522]: DEBUG nova.virt.hardware [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 988.972134] env[65522]: DEBUG nova.virt.hardware [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 988.972134] env[65522]: DEBUG nova.virt.hardware [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 988.972134] env[65522]: DEBUG nova.virt.hardware [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 988.972134] env[65522]: DEBUG nova.virt.hardware [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 988.972134] env[65522]: DEBUG nova.virt.hardware [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 988.972134] env[65522]: DEBUG nova.virt.hardware [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 988.972541] env[65522]: DEBUG nova.virt.hardware [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 988.973585] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-674af796-a455-4115-a00d-ba380e79536d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.977274] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 988.988038] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bbd4b40-2519-4def-aeae-1eaeb91c1fbc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.007219] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 989.012848] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 989.015779] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 989.016230] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b2ec73b9-0980-4f76-97e5-290ae7b3cbc6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.037156] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 989.037156] env[65522]: value = "task-5114505" [ 989.037156] env[65522]: _type = "Task" [ 989.037156] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.042224] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a90e10-2cd5-4d08-8050-6dd6eeeb593a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.054020] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114505, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.054020] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f3ac466-3b21-48ad-9808-37c9b5d52d8d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.086887] env[65522]: DEBUG nova.network.neutron [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Successfully created port: e97ec63d-d242-4026-a95b-36edcd36e5e7 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 989.090425] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0bff92f-05df-4c93-8a15-c569e2ddc276 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.099451] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf00c8b0-36f6-4cd8-b4c1-50a0df48ced3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.116366] env[65522]: DEBUG nova.compute.provider_tree [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.204663] env[65522]: DEBUG nova.compute.manager [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 989.549300] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114505, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.618641] env[65522]: DEBUG nova.scheduler.client.report [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 990.049197] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114505, 'name': CreateVM_Task, 'duration_secs': 0.57981} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.049404] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 990.049848] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.050017] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 990.050367] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 990.050682] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30090892-a58e-4c08-af46-b5214bceaab3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.056766] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 990.056766] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e1471d-f8bf-2a88-9034-e7c8f25375fd" [ 990.056766] env[65522]: _type = "Task" [ 990.056766] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.065568] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e1471d-f8bf-2a88-9034-e7c8f25375fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.124782] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.928s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.127468] env[65522]: DEBUG oslo_concurrency.lockutils [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.536s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 990.128854] env[65522]: DEBUG nova.objects.instance [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lazy-loading 'resources' on Instance uuid 84d4c689-b437-41a5-bb62-d144a30c6219 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.151287] env[65522]: INFO nova.scheduler.client.report [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Deleted allocations for instance 68d8d59b-4fc8-4993-8c9e-31b2329c88dd [ 990.215572] env[65522]: DEBUG nova.compute.manager [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 990.245838] env[65522]: DEBUG nova.virt.hardware [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 990.246204] env[65522]: DEBUG nova.virt.hardware [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 990.246341] env[65522]: DEBUG nova.virt.hardware [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 990.246535] env[65522]: DEBUG nova.virt.hardware [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 990.246682] env[65522]: DEBUG nova.virt.hardware [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 990.246852] env[65522]: DEBUG nova.virt.hardware [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 990.247123] env[65522]: DEBUG nova.virt.hardware [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 990.247911] env[65522]: DEBUG nova.virt.hardware [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 990.247911] env[65522]: DEBUG nova.virt.hardware [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 990.247911] env[65522]: DEBUG nova.virt.hardware [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 990.247911] env[65522]: DEBUG nova.virt.hardware [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 990.248746] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e5b15d1-d410-47d7-9984-82e22a692f36 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.258925] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b45846-3d41-4c69-9b7c-0f13c48f941c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.568496] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e1471d-f8bf-2a88-9034-e7c8f25375fd, 'name': SearchDatastore_Task, 'duration_secs': 0.011543} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.568496] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 990.568496] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.568496] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.568496] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 990.568997] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.568997] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2cf7990b-2f61-467f-9772-97c43a6e09ff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.580488] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.580743] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 990.581640] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c813074c-ce27-40a0-b047-dc534b5a37f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.590781] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 990.590781] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e9571b-d060-9d8f-fbce-c9acc9a4299c" [ 990.590781] env[65522]: _type = "Task" [ 990.590781] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.590781] env[65522]: DEBUG nova.compute.manager [req-32fa2ae3-e68a-46d4-b668-9a7929e4308e req-ea739937-809e-4921-87c7-3abdc2d27c51 service nova] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Received event network-vif-plugged-e97ec63d-d242-4026-a95b-36edcd36e5e7 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 990.590781] env[65522]: DEBUG oslo_concurrency.lockutils [req-32fa2ae3-e68a-46d4-b668-9a7929e4308e req-ea739937-809e-4921-87c7-3abdc2d27c51 service nova] Acquiring lock "00fcbbcb-ef87-4318-8c6e-ce62feb9fd83-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 990.590781] env[65522]: DEBUG oslo_concurrency.lockutils [req-32fa2ae3-e68a-46d4-b668-9a7929e4308e req-ea739937-809e-4921-87c7-3abdc2d27c51 service nova] Lock "00fcbbcb-ef87-4318-8c6e-ce62feb9fd83-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 990.591017] env[65522]: DEBUG oslo_concurrency.lockutils [req-32fa2ae3-e68a-46d4-b668-9a7929e4308e req-ea739937-809e-4921-87c7-3abdc2d27c51 service nova] Lock "00fcbbcb-ef87-4318-8c6e-ce62feb9fd83-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.591054] env[65522]: DEBUG nova.compute.manager [req-32fa2ae3-e68a-46d4-b668-9a7929e4308e req-ea739937-809e-4921-87c7-3abdc2d27c51 service nova] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] No waiting events found dispatching network-vif-plugged-e97ec63d-d242-4026-a95b-36edcd36e5e7 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 990.591219] env[65522]: WARNING nova.compute.manager [req-32fa2ae3-e68a-46d4-b668-9a7929e4308e req-ea739937-809e-4921-87c7-3abdc2d27c51 service nova] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Received unexpected event network-vif-plugged-e97ec63d-d242-4026-a95b-36edcd36e5e7 for instance with vm_state building and task_state spawning. [ 990.601176] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e9571b-d060-9d8f-fbce-c9acc9a4299c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.659397] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d225d33-bc8a-4e9d-a8e0-1f513f4eeab1 tempest-ServersV294TestFqdnHostnames-1853327399 tempest-ServersV294TestFqdnHostnames-1853327399-project-member] Lock "68d8d59b-4fc8-4993-8c9e-31b2329c88dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.234s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.673876] env[65522]: DEBUG nova.network.neutron [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Successfully updated port: e97ec63d-d242-4026-a95b-36edcd36e5e7 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 990.926795] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a8355ce-a5a9-44d0-8cd6-ef5111aa4445 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.937716] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52750f00-72a9-493b-ba9d-08d3a4323a29 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.972547] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57af3a01-3049-4913-8904-816650287e44 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.982455] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abbbbada-0263-421f-95ab-88024530b4da {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.998895] env[65522]: DEBUG nova.compute.provider_tree [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.101949] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e9571b-d060-9d8f-fbce-c9acc9a4299c, 'name': SearchDatastore_Task, 'duration_secs': 0.013183} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.102616] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-febb5fbe-3436-4552-a177-2d688f63cbe9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.108516] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 991.108516] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a29cde-ce2e-1e0a-4126-38c9bfd4a537" [ 991.108516] env[65522]: _type = "Task" [ 991.108516] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.118030] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a29cde-ce2e-1e0a-4126-38c9bfd4a537, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.180082] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Acquiring lock "refresh_cache-00fcbbcb-ef87-4318-8c6e-ce62feb9fd83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.180194] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Acquired lock "refresh_cache-00fcbbcb-ef87-4318-8c6e-ce62feb9fd83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 991.180386] env[65522]: DEBUG nova.network.neutron [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 991.502059] env[65522]: DEBUG nova.scheduler.client.report [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 991.620616] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a29cde-ce2e-1e0a-4126-38c9bfd4a537, 'name': SearchDatastore_Task, 'duration_secs': 0.012374} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.620991] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 991.621256] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 7701b92d-805e-4837-b900-326910b5eef5/7701b92d-805e-4837-b900-326910b5eef5.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 991.621546] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9a3859d-6c8f-42a1-a6cc-7232d5d0624f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.629255] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 991.629255] env[65522]: value = "task-5114506" [ 991.629255] env[65522]: _type = "Task" [ 991.629255] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.640524] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114506, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.683331] env[65522]: WARNING openstack [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.683723] env[65522]: WARNING openstack [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.738054] env[65522]: DEBUG nova.network.neutron [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 991.781588] env[65522]: WARNING openstack [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.782018] env[65522]: WARNING openstack [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.879956] env[65522]: WARNING neutronclient.v2_0.client [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 991.882358] env[65522]: WARNING openstack [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.882358] env[65522]: WARNING openstack [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 992.012874] env[65522]: DEBUG oslo_concurrency.lockutils [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.884s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 992.018955] env[65522]: DEBUG oslo_concurrency.lockutils [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.779s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 992.018955] env[65522]: DEBUG nova.objects.instance [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lazy-loading 'resources' on Instance uuid bb20d99f-fc49-47bf-a6b7-74ea966b71eb {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.053288] env[65522]: INFO nova.scheduler.client.report [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleted allocations for instance 84d4c689-b437-41a5-bb62-d144a30c6219 [ 992.085980] env[65522]: DEBUG nova.network.neutron [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Updating instance_info_cache with network_info: [{"id": "e97ec63d-d242-4026-a95b-36edcd36e5e7", "address": "fa:16:3e:74:dd:d0", "network": {"id": "eff6c543-9811-44f2-a2fc-545e7b475ce0", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-2110660309-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "535709df1c2542f9a9d486d7153af6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape97ec63d-d2", "ovs_interfaceid": "e97ec63d-d242-4026-a95b-36edcd36e5e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 992.143680] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114506, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507182} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.143680] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 7701b92d-805e-4837-b900-326910b5eef5/7701b92d-805e-4837-b900-326910b5eef5.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 992.143680] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 992.143680] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-774ed651-bb67-4fb9-899f-91aa32429577 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.153285] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 992.153285] env[65522]: value = "task-5114507" [ 992.153285] env[65522]: _type = "Task" [ 992.153285] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.165251] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114507, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.563457] env[65522]: DEBUG oslo_concurrency.lockutils [None req-35a5c6e6-37c0-4514-acbc-5e2045b36a0c tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "84d4c689-b437-41a5-bb62-d144a30c6219" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.249s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 992.591430] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Releasing lock "refresh_cache-00fcbbcb-ef87-4318-8c6e-ce62feb9fd83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 992.591430] env[65522]: DEBUG nova.compute.manager [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Instance network_info: |[{"id": "e97ec63d-d242-4026-a95b-36edcd36e5e7", "address": "fa:16:3e:74:dd:d0", "network": {"id": "eff6c543-9811-44f2-a2fc-545e7b475ce0", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-2110660309-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "535709df1c2542f9a9d486d7153af6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape97ec63d-d2", "ovs_interfaceid": "e97ec63d-d242-4026-a95b-36edcd36e5e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 992.591430] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:dd:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9ec24851-7bb6-426b-b28f-f7b246df1713', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e97ec63d-d242-4026-a95b-36edcd36e5e7', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 992.599382] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Creating folder: Project (535709df1c2542f9a9d486d7153af6ba). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 992.603146] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-50753c03-94a3-402a-a61a-d2b16fceeaf2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.617944] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Created folder: Project (535709df1c2542f9a9d486d7153af6ba) in parent group-v994660. [ 992.618410] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Creating folder: Instances. Parent ref: group-v994877. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 992.624356] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d565d3b-4a0d-4a1c-a413-11291216e986 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.626660] env[65522]: DEBUG nova.compute.manager [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Received event network-changed-e97ec63d-d242-4026-a95b-36edcd36e5e7 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 992.626856] env[65522]: DEBUG nova.compute.manager [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Refreshing instance network info cache due to event network-changed-e97ec63d-d242-4026-a95b-36edcd36e5e7. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 992.627092] env[65522]: DEBUG oslo_concurrency.lockutils [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] Acquiring lock "refresh_cache-00fcbbcb-ef87-4318-8c6e-ce62feb9fd83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.627237] env[65522]: DEBUG oslo_concurrency.lockutils [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] Acquired lock "refresh_cache-00fcbbcb-ef87-4318-8c6e-ce62feb9fd83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 992.627392] env[65522]: DEBUG nova.network.neutron [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Refreshing network info cache for port e97ec63d-d242-4026-a95b-36edcd36e5e7 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 992.643858] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Created folder: Instances in parent group-v994877. [ 992.644037] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 992.644234] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 992.644462] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-97360555-0caf-402c-ab8c-c1d2b3ece9f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.675976] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114507, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086294} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.677973] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 992.678142] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 992.678142] env[65522]: value = "task-5114510" [ 992.678142] env[65522]: _type = "Task" [ 992.678142] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.681889] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b81bcf9a-1fd2-49d3-8402-e1edd5079e44 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.708177] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 7701b92d-805e-4837-b900-326910b5eef5/7701b92d-805e-4837-b900-326910b5eef5.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.715908] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b470d9b4-4648-4bef-bf31-4dc5985d3529 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.732992] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114510, 'name': CreateVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.740164] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 992.740164] env[65522]: value = "task-5114511" [ 992.740164] env[65522]: _type = "Task" [ 992.740164] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.758467] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114511, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.961656] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff5a4e5-0a45-4f27-b4c1-391af8d3efc8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.973260] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faaaec25-27b3-4b30-a63b-490cc1ef38aa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.010464] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef3cf7e-3554-44af-be95-73b1257b6eac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.019202] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de8f240-cddd-4497-9859-5c25d017da5a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.035541] env[65522]: DEBUG nova.compute.provider_tree [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.131785] env[65522]: WARNING neutronclient.v2_0.client [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 993.132617] env[65522]: WARNING openstack [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.133107] env[65522]: WARNING openstack [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.197150] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114510, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.256847] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114511, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.304891] env[65522]: WARNING openstack [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.306429] env[65522]: WARNING openstack [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.411347] env[65522]: WARNING neutronclient.v2_0.client [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 993.412032] env[65522]: WARNING openstack [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.412397] env[65522]: WARNING openstack [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.518403] env[65522]: DEBUG nova.network.neutron [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Updated VIF entry in instance network info cache for port e97ec63d-d242-4026-a95b-36edcd36e5e7. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 993.518814] env[65522]: DEBUG nova.network.neutron [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Updating instance_info_cache with network_info: [{"id": "e97ec63d-d242-4026-a95b-36edcd36e5e7", "address": "fa:16:3e:74:dd:d0", "network": {"id": "eff6c543-9811-44f2-a2fc-545e7b475ce0", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-2110660309-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "535709df1c2542f9a9d486d7153af6ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9ec24851-7bb6-426b-b28f-f7b246df1713", "external-id": "nsx-vlan-transportzone-359", "segmentation_id": 359, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape97ec63d-d2", "ovs_interfaceid": "e97ec63d-d242-4026-a95b-36edcd36e5e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 993.540047] env[65522]: DEBUG nova.scheduler.client.report [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 993.697780] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114510, 'name': CreateVM_Task, 'duration_secs': 0.713894} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.698173] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 993.698520] env[65522]: WARNING neutronclient.v2_0.client [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 993.698898] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.699071] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 993.699389] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 993.699666] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f861521-eef5-4ae5-bb37-00649660f072 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.705395] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Waiting for the task: (returnval){ [ 993.705395] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52613ab5-b65f-0f98-5d3f-ba3d7f50404d" [ 993.705395] env[65522]: _type = "Task" [ 993.705395] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.716461] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52613ab5-b65f-0f98-5d3f-ba3d7f50404d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.751567] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114511, 'name': ReconfigVM_Task, 'duration_secs': 0.657169} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.751892] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 7701b92d-805e-4837-b900-326910b5eef5/7701b92d-805e-4837-b900-326910b5eef5.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.752584] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c4c3ecb5-c62d-402b-b5a4-7da7b6bfff2f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.760251] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 993.760251] env[65522]: value = "task-5114512" [ 993.760251] env[65522]: _type = "Task" [ 993.760251] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.770349] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114512, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.022592] env[65522]: DEBUG oslo_concurrency.lockutils [req-b9cdd757-6cd7-4df0-a607-b66ad32c43cb req-d55d9ae4-bd62-4f04-b4ea-c221395749dc service nova] Releasing lock "refresh_cache-00fcbbcb-ef87-4318-8c6e-ce62feb9fd83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 994.045872] env[65522]: DEBUG oslo_concurrency.lockutils [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.028s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 994.048493] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.925s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 994.048745] env[65522]: DEBUG nova.objects.instance [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lazy-loading 'resources' on Instance uuid 82a7ad85-a061-4eec-bd3f-fc977532dfbf {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 994.074634] env[65522]: INFO nova.scheduler.client.report [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Deleted allocations for instance bb20d99f-fc49-47bf-a6b7-74ea966b71eb [ 994.217921] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52613ab5-b65f-0f98-5d3f-ba3d7f50404d, 'name': SearchDatastore_Task, 'duration_secs': 0.030413} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.218271] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 994.218517] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 994.218757] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.218900] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 994.219087] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 994.219373] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-497cc35b-00e5-45b0-a259-096ab0f85057 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.230478] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 994.230804] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 994.232040] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50f27573-96f7-437c-83ff-b283b60d04db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.238625] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Waiting for the task: (returnval){ [ 994.238625] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5231865d-f8ea-6ded-6e98-4d2546105fc4" [ 994.238625] env[65522]: _type = "Task" [ 994.238625] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.248101] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5231865d-f8ea-6ded-6e98-4d2546105fc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.271425] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114512, 'name': Rename_Task, 'duration_secs': 0.21905} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.271592] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 994.271821] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-56eda48a-8006-4bb4-a3e7-206d97a13983 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.279588] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 994.279588] env[65522]: value = "task-5114513" [ 994.279588] env[65522]: _type = "Task" [ 994.279588] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.289152] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114513, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.585609] env[65522]: DEBUG oslo_concurrency.lockutils [None req-46c041a7-d80c-4caa-8751-de38932084f4 tempest-ServersAdminTestJSON-1076150104 tempest-ServersAdminTestJSON-1076150104-project-member] Lock "bb20d99f-fc49-47bf-a6b7-74ea966b71eb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.810s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 994.750255] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5231865d-f8ea-6ded-6e98-4d2546105fc4, 'name': SearchDatastore_Task, 'duration_secs': 0.017038} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.753907] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-812233e8-1520-4878-868f-f27edd419fb8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.760918] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Waiting for the task: (returnval){ [ 994.760918] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529c031d-5dcc-3f91-0037-f44ed0c7ccdf" [ 994.760918] env[65522]: _type = "Task" [ 994.760918] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.772259] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529c031d-5dcc-3f91-0037-f44ed0c7ccdf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.790226] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114513, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.816095] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16955585-ef12-4b70-9790-f027e939714b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.825598] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41887d0a-60a0-4697-a509-6ab3eb1dd67a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.859043] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29eb1bb0-0473-476b-9933-cf71d8dda7d0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.870040] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23056758-465c-4ac6-9be5-333ed02eed6b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.886570] env[65522]: DEBUG nova.compute.provider_tree [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.074772] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "99353848-2f0f-4388-9fcd-91e799342386" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 995.074772] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "99353848-2f0f-4388-9fcd-91e799342386" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.272446] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529c031d-5dcc-3f91-0037-f44ed0c7ccdf, 'name': SearchDatastore_Task, 'duration_secs': 0.016386} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.272745] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 995.273048] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83/00fcbbcb-ef87-4318-8c6e-ce62feb9fd83.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 995.273353] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0a39256-3090-4bbb-b5d3-371941d15470 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.281954] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Waiting for the task: (returnval){ [ 995.281954] env[65522]: value = "task-5114514" [ 995.281954] env[65522]: _type = "Task" [ 995.281954] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.296150] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114514, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.299167] env[65522]: DEBUG oslo_vmware.api [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114513, 'name': PowerOnVM_Task, 'duration_secs': 0.633424} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.299600] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.299905] env[65522]: DEBUG nova.compute.manager [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 995.301519] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d70511-3ff2-4d6f-923c-e9012b4a3b71 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.391272] env[65522]: DEBUG nova.scheduler.client.report [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 995.579022] env[65522]: DEBUG nova.compute.manager [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 995.804091] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114514, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.824283] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 995.896521] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.848s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 995.899095] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.241s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.899278] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 995.899428] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 995.899806] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.759s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.901293] env[65522]: INFO nova.compute.claims [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 995.910130] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16c2c9f-fe55-4fd0-9925-6025d9776d41 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.920054] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc13591-c58a-439c-9766-442c1ecb0e72 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.936225] env[65522]: INFO nova.scheduler.client.report [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleted allocations for instance 82a7ad85-a061-4eec-bd3f-fc977532dfbf [ 995.938371] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b567e6-2580-4e8b-8aee-15110f2e9569 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.950022] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799c5f3b-34ee-46de-bac5-031566a38360 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.981264] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=177609MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 995.981428] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 996.105501] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 996.300304] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114514, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.593951} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.300923] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83/00fcbbcb-ef87-4318-8c6e-ce62feb9fd83.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 996.300923] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 996.301169] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b01d2816-2056-4a14-a435-1fe9051273af {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.309748] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Waiting for the task: (returnval){ [ 996.309748] env[65522]: value = "task-5114515" [ 996.309748] env[65522]: _type = "Task" [ 996.309748] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.320760] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114515, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.447641] env[65522]: DEBUG oslo_concurrency.lockutils [None req-31f1294d-b28b-4d35-8c6c-276b528ceaf1 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "82a7ad85-a061-4eec-bd3f-fc977532dfbf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.813s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 996.822029] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114515, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.32509} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.822361] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 996.824032] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f25363-b9fe-4358-910e-ea45f384478d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.855739] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83/00fcbbcb-ef87-4318-8c6e-ce62feb9fd83.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 996.856198] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e6a564b-6919-4288-baf2-47659372539f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.878324] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Waiting for the task: (returnval){ [ 996.878324] env[65522]: value = "task-5114516" [ 996.878324] env[65522]: _type = "Task" [ 996.878324] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.887816] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114516, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.027143] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "7701b92d-805e-4837-b900-326910b5eef5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 997.027413] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "7701b92d-805e-4837-b900-326910b5eef5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.027624] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "7701b92d-805e-4837-b900-326910b5eef5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 997.027804] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "7701b92d-805e-4837-b900-326910b5eef5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.027970] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "7701b92d-805e-4837-b900-326910b5eef5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 997.032482] env[65522]: INFO nova.compute.manager [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Terminating instance [ 997.276667] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7806385f-5e22-4adf-ac3b-94e1cd3b6404 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.285675] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f76365-9db7-4802-bcb0-c5e879467f0d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.325543] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd767d9e-0daf-4fff-918e-ad3c0d1ee4f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.334285] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2721482-c503-40a3-9ace-f7edeb9c9e40 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.350279] env[65522]: DEBUG nova.compute.provider_tree [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.391024] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114516, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.543889] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "refresh_cache-7701b92d-805e-4837-b900-326910b5eef5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.543983] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquired lock "refresh_cache-7701b92d-805e-4837-b900-326910b5eef5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 997.545142] env[65522]: DEBUG nova.network.neutron [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 997.853649] env[65522]: DEBUG nova.scheduler.client.report [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 997.890359] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114516, 'name': ReconfigVM_Task, 'duration_secs': 0.673772} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.890651] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83/00fcbbcb-ef87-4318-8c6e-ce62feb9fd83.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.891408] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0191db7d-80b9-4641-a081-a57d6acef706 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.067709] env[65522]: WARNING neutronclient.v2_0.client [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 998.070437] env[65522]: WARNING openstack [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 998.070437] env[65522]: WARNING openstack [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 998.086416] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Waiting for the task: (returnval){ [ 998.086416] env[65522]: value = "task-5114517" [ 998.086416] env[65522]: _type = "Task" [ 998.086416] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.102234] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114517, 'name': Rename_Task, 'duration_secs': 0.175058} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.103256] env[65522]: DEBUG nova.network.neutron [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 998.105125] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 998.106589] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0956c7c4-bdca-475b-b425-a4a1cac6dd13 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.115296] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Waiting for the task: (returnval){ [ 998.115296] env[65522]: value = "task-5114518" [ 998.115296] env[65522]: _type = "Task" [ 998.115296] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.126703] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114518, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.194800] env[65522]: DEBUG nova.network.neutron [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 998.280429] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "7a358133-88a5-4dd7-ab69-45b8172870a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.280429] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "7a358133-88a5-4dd7-ab69-45b8172870a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 998.359159] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 998.359838] env[65522]: DEBUG nova.compute.manager [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 998.366636] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.723s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 998.367736] env[65522]: INFO nova.compute.claims [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 998.632642] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114518, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.697146] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Releasing lock "refresh_cache-7701b92d-805e-4837-b900-326910b5eef5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 998.697615] env[65522]: DEBUG nova.compute.manager [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 998.697819] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 998.698766] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95af628-7cf6-4414-89c6-c6477c78b185 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.707420] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 998.707485] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31c0f551-7d90-4a3a-aa5a-4b9ea0b95d6e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.715015] env[65522]: DEBUG oslo_vmware.api [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 998.715015] env[65522]: value = "task-5114519" [ 998.715015] env[65522]: _type = "Task" [ 998.715015] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.726232] env[65522]: DEBUG oslo_vmware.api [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114519, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.783705] env[65522]: DEBUG nova.compute.manager [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 998.875205] env[65522]: DEBUG nova.compute.utils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 998.879829] env[65522]: DEBUG nova.compute.manager [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 998.880065] env[65522]: DEBUG nova.network.neutron [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 998.880427] env[65522]: WARNING neutronclient.v2_0.client [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 998.880806] env[65522]: WARNING neutronclient.v2_0.client [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 998.881455] env[65522]: WARNING openstack [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 998.882749] env[65522]: WARNING openstack [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 998.953412] env[65522]: DEBUG nova.policy [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f51006be1b2e401a8dd5504e0e481810', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ea0fab7de6c47eeaf3bdde413ae3901', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 999.129396] env[65522]: DEBUG oslo_vmware.api [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114518, 'name': PowerOnVM_Task, 'duration_secs': 0.813858} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.129679] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 999.129946] env[65522]: INFO nova.compute.manager [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Took 8.91 seconds to spawn the instance on the hypervisor. [ 999.130157] env[65522]: DEBUG nova.compute.manager [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 999.131012] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6ecab9-0e19-437e-953a-3a9ed46e1ee7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.229540] env[65522]: DEBUG oslo_vmware.api [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114519, 'name': PowerOffVM_Task, 'duration_secs': 0.280585} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.230244] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 999.230244] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 999.230634] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff426ef9-3cc4-4b84-a0d4-4979deffca99 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.267241] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 999.267472] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 999.267568] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Deleting the datastore file [datastore1] 7701b92d-805e-4837-b900-326910b5eef5 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 999.267893] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e2b8f49c-6a30-4feb-b36d-5066cdb4fc3a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.276826] env[65522]: DEBUG oslo_vmware.api [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 999.276826] env[65522]: value = "task-5114521" [ 999.276826] env[65522]: _type = "Task" [ 999.276826] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.290493] env[65522]: DEBUG oslo_vmware.api [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114521, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.309199] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 999.329096] env[65522]: DEBUG nova.network.neutron [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Successfully created port: cc030045-9833-4322-9675-37b3f0f31161 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 999.380441] env[65522]: DEBUG nova.compute.manager [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 999.653086] env[65522]: INFO nova.compute.manager [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Took 28.37 seconds to build instance. [ 999.712413] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235925ca-dbff-4328-8c95-2704db1b72b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.720958] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aabf96f-6482-455b-a0f7-38e6fa4c0071 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.756043] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9e11e7-a124-435c-b28b-b2222548bc7b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.764527] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e18202-c909-429b-877d-84ad4d5a8218 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.781046] env[65522]: DEBUG nova.compute.provider_tree [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.791265] env[65522]: DEBUG oslo_vmware.api [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114521, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.120072} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.791640] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.792249] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 999.794759] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 999.794759] env[65522]: INFO nova.compute.manager [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Took 1.09 seconds to destroy the instance on the hypervisor. [ 999.794759] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 999.794759] env[65522]: DEBUG nova.compute.manager [-] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 999.794759] env[65522]: DEBUG nova.network.neutron [-] [instance: 7701b92d-805e-4837-b900-326910b5eef5] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 999.794759] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 999.795205] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.795771] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.819362] env[65522]: DEBUG nova.network.neutron [-] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 999.819615] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1000.156585] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69d10336-352a-44ba-b874-047d7989d4f8 tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Lock "00fcbbcb-ef87-4318-8c6e-ce62feb9fd83" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.882s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.182401] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Acquiring lock "00fcbbcb-ef87-4318-8c6e-ce62feb9fd83" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.182816] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Lock "00fcbbcb-ef87-4318-8c6e-ce62feb9fd83" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.183335] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Acquiring lock "00fcbbcb-ef87-4318-8c6e-ce62feb9fd83-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.183462] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Lock "00fcbbcb-ef87-4318-8c6e-ce62feb9fd83-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.183648] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Lock "00fcbbcb-ef87-4318-8c6e-ce62feb9fd83-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.188835] env[65522]: INFO nova.compute.manager [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Terminating instance [ 1000.286824] env[65522]: DEBUG nova.scheduler.client.report [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1000.321882] env[65522]: DEBUG nova.network.neutron [-] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1000.390203] env[65522]: DEBUG nova.compute.manager [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1000.418279] env[65522]: DEBUG nova.virt.hardware [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1000.418609] env[65522]: DEBUG nova.virt.hardware [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1000.418810] env[65522]: DEBUG nova.virt.hardware [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1000.419017] env[65522]: DEBUG nova.virt.hardware [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1000.419161] env[65522]: DEBUG nova.virt.hardware [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1000.419302] env[65522]: DEBUG nova.virt.hardware [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1000.419509] env[65522]: DEBUG nova.virt.hardware [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1000.419662] env[65522]: DEBUG nova.virt.hardware [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1000.419823] env[65522]: DEBUG nova.virt.hardware [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1000.419981] env[65522]: DEBUG nova.virt.hardware [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1000.420259] env[65522]: DEBUG nova.virt.hardware [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1000.421498] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34296b8-f71b-4250-8e22-0d15c6033660 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.430132] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13993b6a-03ed-4cbf-b9b6-42e982d3621a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.517013] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquiring lock "b412d3e5-0536-413c-9519-5f85c7647cdf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.517260] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Lock "b412d3e5-0536-413c-9519-5f85c7647cdf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.692675] env[65522]: DEBUG nova.compute.manager [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1000.693082] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.694273] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d742f8-6a83-4851-a3e7-7639e798b3ad {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.704203] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.704430] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-917b56b0-c88e-45bc-aa84-2f3055e7118d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.711518] env[65522]: DEBUG oslo_vmware.api [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Waiting for the task: (returnval){ [ 1000.711518] env[65522]: value = "task-5114522" [ 1000.711518] env[65522]: _type = "Task" [ 1000.711518] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.720864] env[65522]: DEBUG oslo_vmware.api [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114522, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.783250] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.783559] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.783755] env[65522]: INFO nova.compute.manager [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Shelving [ 1000.792372] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.426s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.792874] env[65522]: DEBUG nova.compute.manager [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1000.796388] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.205s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.796640] env[65522]: DEBUG nova.objects.instance [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lazy-loading 'resources' on Instance uuid 8da2a500-6f0e-4eda-9a92-79510753b3b8 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.799170] env[65522]: DEBUG nova.compute.manager [req-84183ce9-16bb-4cc4-9efb-30f92ee70e49 req-d3b57905-e445-4942-94ca-4775305fcb61 service nova] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Received event network-vif-plugged-cc030045-9833-4322-9675-37b3f0f31161 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1000.799170] env[65522]: DEBUG oslo_concurrency.lockutils [req-84183ce9-16bb-4cc4-9efb-30f92ee70e49 req-d3b57905-e445-4942-94ca-4775305fcb61 service nova] Acquiring lock "8859b051-8f75-4aad-b789-42662019d4c5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.799371] env[65522]: DEBUG oslo_concurrency.lockutils [req-84183ce9-16bb-4cc4-9efb-30f92ee70e49 req-d3b57905-e445-4942-94ca-4775305fcb61 service nova] Lock "8859b051-8f75-4aad-b789-42662019d4c5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.799676] env[65522]: DEBUG oslo_concurrency.lockutils [req-84183ce9-16bb-4cc4-9efb-30f92ee70e49 req-d3b57905-e445-4942-94ca-4775305fcb61 service nova] Lock "8859b051-8f75-4aad-b789-42662019d4c5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.799676] env[65522]: DEBUG nova.compute.manager [req-84183ce9-16bb-4cc4-9efb-30f92ee70e49 req-d3b57905-e445-4942-94ca-4775305fcb61 service nova] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] No waiting events found dispatching network-vif-plugged-cc030045-9833-4322-9675-37b3f0f31161 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1000.799939] env[65522]: WARNING nova.compute.manager [req-84183ce9-16bb-4cc4-9efb-30f92ee70e49 req-d3b57905-e445-4942-94ca-4775305fcb61 service nova] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Received unexpected event network-vif-plugged-cc030045-9833-4322-9675-37b3f0f31161 for instance with vm_state building and task_state spawning. [ 1000.824609] env[65522]: INFO nova.compute.manager [-] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Took 1.03 seconds to deallocate network for instance. [ 1000.885776] env[65522]: DEBUG nova.network.neutron [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Successfully updated port: cc030045-9833-4322-9675-37b3f0f31161 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1001.020129] env[65522]: DEBUG nova.compute.manager [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1001.228913] env[65522]: DEBUG oslo_vmware.api [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114522, 'name': PowerOffVM_Task, 'duration_secs': 0.195873} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.229398] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.229398] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1001.229634] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-120a9a78-f0bd-4d96-9cb4-b67a8a505784 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.303171] env[65522]: DEBUG nova.compute.utils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1001.308242] env[65522]: DEBUG nova.compute.manager [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1001.308560] env[65522]: DEBUG nova.network.neutron [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1001.308973] env[65522]: WARNING neutronclient.v2_0.client [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1001.309333] env[65522]: WARNING neutronclient.v2_0.client [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1001.309999] env[65522]: WARNING openstack [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.310366] env[65522]: WARNING openstack [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.320630] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.320883] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.321083] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Deleting the datastore file [datastore1] 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.321604] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7e223031-c8f6-488b-b100-bab4cfb68162 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.334040] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1001.334413] env[65522]: DEBUG oslo_vmware.api [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Waiting for the task: (returnval){ [ 1001.334413] env[65522]: value = "task-5114524" [ 1001.334413] env[65522]: _type = "Task" [ 1001.334413] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.344335] env[65522]: DEBUG oslo_vmware.api [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114524, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.377120] env[65522]: DEBUG nova.policy [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b51f1b47ad744dc2b2b0598cee11bc6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d1239b79ae94cceb89ae7a8bd57da08', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1001.389206] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "refresh_cache-8859b051-8f75-4aad-b789-42662019d4c5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.389356] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "refresh_cache-8859b051-8f75-4aad-b789-42662019d4c5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1001.389459] env[65522]: DEBUG nova.network.neutron [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1001.543517] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1001.630941] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745dc24a-f247-4802-9117-579b763eb803 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.640491] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b20f55-92c8-48ff-89b0-d4d5bd387ef2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.679996] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86310ac4-99f4-4bd1-8a9e-6fc9dcd837c0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.690395] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9847b9-82fe-4fde-9fa1-6fa9cdf209a6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.706446] env[65522]: DEBUG nova.network.neutron [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Successfully created port: ed6e9224-69f8-46fd-a71f-27d8bc6eaec3 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1001.709168] env[65522]: DEBUG nova.compute.provider_tree [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.795397] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.795827] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a27a22a7-f566-4d28-81d8-b3c465817e92 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.803807] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1001.803807] env[65522]: value = "task-5114525" [ 1001.803807] env[65522]: _type = "Task" [ 1001.803807] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.816091] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114525, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.819039] env[65522]: DEBUG nova.compute.manager [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1001.848093] env[65522]: DEBUG oslo_vmware.api [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114524, 'name': DeleteDatastoreFile_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.898411] env[65522]: WARNING openstack [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.898916] env[65522]: WARNING openstack [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.937941] env[65522]: DEBUG nova.network.neutron [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1001.958585] env[65522]: WARNING openstack [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.958993] env[65522]: WARNING openstack [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1002.101669] env[65522]: WARNING neutronclient.v2_0.client [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1002.102645] env[65522]: WARNING openstack [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1002.103123] env[65522]: WARNING openstack [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1002.212691] env[65522]: DEBUG nova.scheduler.client.report [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1002.235312] env[65522]: DEBUG nova.network.neutron [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Updating instance_info_cache with network_info: [{"id": "cc030045-9833-4322-9675-37b3f0f31161", "address": "fa:16:3e:fe:83:f5", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc030045-98", "ovs_interfaceid": "cc030045-9833-4322-9675-37b3f0f31161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1002.313989] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114525, 'name': PowerOffVM_Task, 'duration_secs': 0.254833} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.314269] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1002.315097] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15f2535-7a43-4152-be86-d4986acbfbe4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.337947] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f89d0260-160a-4827-bd0e-1cc83e77a892 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.352915] env[65522]: DEBUG oslo_vmware.api [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Task: {'id': task-5114524, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.516963} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.353189] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.353369] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.353542] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.353708] env[65522]: INFO nova.compute.manager [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Took 1.66 seconds to destroy the instance on the hypervisor. [ 1002.353947] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1002.354156] env[65522]: DEBUG nova.compute.manager [-] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1002.354255] env[65522]: DEBUG nova.network.neutron [-] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1002.354495] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1002.355025] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1002.355286] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1002.402976] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1002.718603] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.922s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1002.721152] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.887s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1002.722785] env[65522]: INFO nova.compute.claims [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1002.739718] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "refresh_cache-8859b051-8f75-4aad-b789-42662019d4c5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1002.740109] env[65522]: DEBUG nova.compute.manager [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Instance network_info: |[{"id": "cc030045-9833-4322-9675-37b3f0f31161", "address": "fa:16:3e:fe:83:f5", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc030045-98", "ovs_interfaceid": "cc030045-9833-4322-9675-37b3f0f31161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1002.740594] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:83:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5e1049e8-c06b-4c93-a9e1-2cbb530f3f95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc030045-9833-4322-9675-37b3f0f31161', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1002.748610] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1002.748863] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1002.749133] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f39bcb30-6c33-45ba-aadb-a82884c6de0c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.765117] env[65522]: INFO nova.scheduler.client.report [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleted allocations for instance 8da2a500-6f0e-4eda-9a92-79510753b3b8 [ 1002.773956] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1002.773956] env[65522]: value = "task-5114526" [ 1002.773956] env[65522]: _type = "Task" [ 1002.773956] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.784385] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114526, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.839259] env[65522]: DEBUG nova.compute.manager [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1002.852161] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1002.852467] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-5b4505ed-4281-4535-bedf-2b82075137ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.863358] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1002.863358] env[65522]: value = "task-5114527" [ 1002.863358] env[65522]: _type = "Task" [ 1002.863358] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.869206] env[65522]: DEBUG nova.virt.hardware [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1002.869447] env[65522]: DEBUG nova.virt.hardware [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1002.869593] env[65522]: DEBUG nova.virt.hardware [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1002.869771] env[65522]: DEBUG nova.virt.hardware [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1002.869913] env[65522]: DEBUG nova.virt.hardware [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1002.870067] env[65522]: DEBUG nova.virt.hardware [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1002.870274] env[65522]: DEBUG nova.virt.hardware [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1002.870428] env[65522]: DEBUG nova.virt.hardware [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1002.870639] env[65522]: DEBUG nova.virt.hardware [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1002.870857] env[65522]: DEBUG nova.virt.hardware [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1002.871053] env[65522]: DEBUG nova.virt.hardware [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1002.871885] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd2dc93f-8402-4703-8e85-c85fcfb58838 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.880295] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114527, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.884516] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521bc3d6-5e46-4c32-a0c4-9490becfc8d8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.910406] env[65522]: DEBUG nova.compute.manager [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Received event network-changed-cc030045-9833-4322-9675-37b3f0f31161 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1002.910590] env[65522]: DEBUG nova.compute.manager [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Refreshing instance network info cache due to event network-changed-cc030045-9833-4322-9675-37b3f0f31161. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1002.910897] env[65522]: DEBUG oslo_concurrency.lockutils [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] Acquiring lock "refresh_cache-8859b051-8f75-4aad-b789-42662019d4c5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.911083] env[65522]: DEBUG oslo_concurrency.lockutils [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] Acquired lock "refresh_cache-8859b051-8f75-4aad-b789-42662019d4c5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1002.911260] env[65522]: DEBUG nova.network.neutron [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Refreshing network info cache for port cc030045-9833-4322-9675-37b3f0f31161 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1003.121500] env[65522]: DEBUG nova.network.neutron [-] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1003.167247] env[65522]: DEBUG nova.compute.manager [req-3bf95a7d-ffa8-4f74-9f69-8dfacfbd5ac8 req-c0bf66bb-b9a5-4d71-83d3-7b03900e274d service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Received event network-vif-plugged-ed6e9224-69f8-46fd-a71f-27d8bc6eaec3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1003.167247] env[65522]: DEBUG oslo_concurrency.lockutils [req-3bf95a7d-ffa8-4f74-9f69-8dfacfbd5ac8 req-c0bf66bb-b9a5-4d71-83d3-7b03900e274d service nova] Acquiring lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1003.167247] env[65522]: DEBUG oslo_concurrency.lockutils [req-3bf95a7d-ffa8-4f74-9f69-8dfacfbd5ac8 req-c0bf66bb-b9a5-4d71-83d3-7b03900e274d service nova] Lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1003.167579] env[65522]: DEBUG oslo_concurrency.lockutils [req-3bf95a7d-ffa8-4f74-9f69-8dfacfbd5ac8 req-c0bf66bb-b9a5-4d71-83d3-7b03900e274d service nova] Lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1003.167680] env[65522]: DEBUG nova.compute.manager [req-3bf95a7d-ffa8-4f74-9f69-8dfacfbd5ac8 req-c0bf66bb-b9a5-4d71-83d3-7b03900e274d service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] No waiting events found dispatching network-vif-plugged-ed6e9224-69f8-46fd-a71f-27d8bc6eaec3 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1003.167867] env[65522]: WARNING nova.compute.manager [req-3bf95a7d-ffa8-4f74-9f69-8dfacfbd5ac8 req-c0bf66bb-b9a5-4d71-83d3-7b03900e274d service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Received unexpected event network-vif-plugged-ed6e9224-69f8-46fd-a71f-27d8bc6eaec3 for instance with vm_state building and task_state spawning. [ 1003.275494] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e2f21073-ee05-4dd1-bc89-c756e2e05f81 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "8da2a500-6f0e-4eda-9a92-79510753b3b8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.055s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1003.289035] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114526, 'name': CreateVM_Task, 'duration_secs': 0.345126} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.289263] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1003.290978] env[65522]: WARNING neutronclient.v2_0.client [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1003.290978] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.290978] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1003.290978] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1003.290978] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-905548da-9a44-4467-95f8-ffba947e1641 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.296686] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1003.296686] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5241c9b2-3fe0-1cf2-492d-e1abf82aa93d" [ 1003.296686] env[65522]: _type = "Task" [ 1003.296686] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.297801] env[65522]: DEBUG nova.network.neutron [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Successfully updated port: ed6e9224-69f8-46fd-a71f-27d8bc6eaec3 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1003.311316] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5241c9b2-3fe0-1cf2-492d-e1abf82aa93d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.374290] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114527, 'name': CreateSnapshot_Task, 'duration_secs': 0.482948} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.374569] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1003.375346] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dcb77ac-303f-4c61-8cbd-34ab0113a948 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.414812] env[65522]: WARNING neutronclient.v2_0.client [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1003.415628] env[65522]: WARNING openstack [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1003.415785] env[65522]: WARNING openstack [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1003.511287] env[65522]: WARNING openstack [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1003.511668] env[65522]: WARNING openstack [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1003.569862] env[65522]: WARNING neutronclient.v2_0.client [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1003.570538] env[65522]: WARNING openstack [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1003.571039] env[65522]: WARNING openstack [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1003.623613] env[65522]: INFO nova.compute.manager [-] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Took 1.27 seconds to deallocate network for instance. [ 1003.652038] env[65522]: DEBUG nova.network.neutron [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Updated VIF entry in instance network info cache for port cc030045-9833-4322-9675-37b3f0f31161. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1003.653030] env[65522]: DEBUG nova.network.neutron [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Updating instance_info_cache with network_info: [{"id": "cc030045-9833-4322-9675-37b3f0f31161", "address": "fa:16:3e:fe:83:f5", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc030045-98", "ovs_interfaceid": "cc030045-9833-4322-9675-37b3f0f31161", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1003.808134] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.808391] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1003.808575] env[65522]: DEBUG nova.network.neutron [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1003.818579] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5241c9b2-3fe0-1cf2-492d-e1abf82aa93d, 'name': SearchDatastore_Task, 'duration_secs': 0.013209} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.821038] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1003.821038] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1003.821038] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.821038] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1003.821038] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1003.821537] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-661bafa8-914c-4327-b2e6-647c7caf37d4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.835484] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1003.836040] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1003.837362] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29dc06ae-e738-415f-9268-981ab9047a2a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.845935] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1003.845935] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cbed32-21b8-d11c-0734-88c243878c3a" [ 1003.845935] env[65522]: _type = "Task" [ 1003.845935] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.857773] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cbed32-21b8-d11c-0734-88c243878c3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.893415] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1003.896505] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-bee639ab-8e02-4cdf-9eb1-762d69178166 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.905981] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1003.905981] env[65522]: value = "task-5114528" [ 1003.905981] env[65522]: _type = "Task" [ 1003.905981] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.915287] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114528, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.022445] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e307c098-1c34-454a-9a9f-33265712f1eb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.031083] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d6a406-5f9d-486d-89e7-cb07509882e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.062503] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-debec6b2-1ce9-45bd-bbb4-29c4d105289c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.070501] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897663f8-0164-41f1-8667-8ec4a7f148c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.085267] env[65522]: DEBUG nova.compute.provider_tree [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.129986] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1004.155264] env[65522]: DEBUG oslo_concurrency.lockutils [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] Releasing lock "refresh_cache-8859b051-8f75-4aad-b789-42662019d4c5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1004.155705] env[65522]: DEBUG nova.compute.manager [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Received event network-vif-deleted-e97ec63d-d242-4026-a95b-36edcd36e5e7 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1004.156075] env[65522]: INFO nova.compute.manager [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Neutron deleted interface e97ec63d-d242-4026-a95b-36edcd36e5e7; detaching it from the instance and deleting it from the info cache [ 1004.156356] env[65522]: DEBUG nova.network.neutron [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1004.311837] env[65522]: WARNING openstack [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1004.312326] env[65522]: WARNING openstack [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1004.354111] env[65522]: DEBUG nova.network.neutron [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1004.359101] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cbed32-21b8-d11c-0734-88c243878c3a, 'name': SearchDatastore_Task, 'duration_secs': 0.010826} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.360036] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-453af663-7a25-4697-94a8-eba1eec78390 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.367033] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1004.367033] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f6ece9-6b4c-854c-7dc0-27a8706a6590" [ 1004.367033] env[65522]: _type = "Task" [ 1004.367033] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.378588] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f6ece9-6b4c-854c-7dc0-27a8706a6590, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.380158] env[65522]: WARNING openstack [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1004.380579] env[65522]: WARNING openstack [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1004.419254] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114528, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.455789] env[65522]: WARNING neutronclient.v2_0.client [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1004.456702] env[65522]: WARNING openstack [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1004.456951] env[65522]: WARNING openstack [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1004.539673] env[65522]: DEBUG nova.network.neutron [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance_info_cache with network_info: [{"id": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "address": "fa:16:3e:c3:12:7d", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped6e9224-69", "ovs_interfaceid": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1004.588326] env[65522]: DEBUG nova.scheduler.client.report [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1004.660550] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e13fdb59-8b4f-497b-b228-6c5e7fabcb6d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.670562] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d64c31-432c-4596-8e71-4749261d40ba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.702355] env[65522]: DEBUG nova.compute.manager [req-7a6616ce-9e8e-47a4-9e84-0b4f44b40a43 req-06536a16-764c-4a07-ad84-ddea9484c3eb service nova] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Detach interface failed, port_id=e97ec63d-d242-4026-a95b-36edcd36e5e7, reason: Instance 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1004.879588] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f6ece9-6b4c-854c-7dc0-27a8706a6590, 'name': SearchDatastore_Task, 'duration_secs': 0.012415} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.879942] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1004.880312] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 8859b051-8f75-4aad-b789-42662019d4c5/8859b051-8f75-4aad-b789-42662019d4c5.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1004.880527] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bdb24c0c-8eaf-4c37-aad5-7ef6d888b1ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.888904] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1004.888904] env[65522]: value = "task-5114529" [ 1004.888904] env[65522]: _type = "Task" [ 1004.888904] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.899950] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114529, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.915846] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114528, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.043036] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1005.043497] env[65522]: DEBUG nova.compute.manager [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Instance network_info: |[{"id": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "address": "fa:16:3e:c3:12:7d", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped6e9224-69", "ovs_interfaceid": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1005.044033] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:12:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a91c3a96-63d0-407c-bcde-c3d5b58d9cb2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ed6e9224-69f8-46fd-a71f-27d8bc6eaec3', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1005.051986] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Creating folder: Project (0d1239b79ae94cceb89ae7a8bd57da08). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1005.052335] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd125722-17a4-48f6-b2dc-274cba73d269 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.067892] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Created folder: Project (0d1239b79ae94cceb89ae7a8bd57da08) in parent group-v994660. [ 1005.068286] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Creating folder: Instances. Parent ref: group-v994883. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1005.068584] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d8ca8eb-20a1-4867-b60e-a180ecca8b32 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.083386] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Created folder: Instances in parent group-v994883. [ 1005.083595] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1005.083818] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1005.084059] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-60931d7a-11f4-4387-aee5-7b19eafa77e2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.100888] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.101447] env[65522]: DEBUG nova.compute.manager [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1005.104592] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.716s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.107069] env[65522]: INFO nova.compute.claims [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.118982] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1005.118982] env[65522]: value = "task-5114532" [ 1005.118982] env[65522]: _type = "Task" [ 1005.118982] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.130208] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114532, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.195586] env[65522]: DEBUG nova.compute.manager [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Received event network-changed-ed6e9224-69f8-46fd-a71f-27d8bc6eaec3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1005.195817] env[65522]: DEBUG nova.compute.manager [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Refreshing instance network info cache due to event network-changed-ed6e9224-69f8-46fd-a71f-27d8bc6eaec3. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1005.196367] env[65522]: DEBUG oslo_concurrency.lockutils [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] Acquiring lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.196562] env[65522]: DEBUG oslo_concurrency.lockutils [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] Acquired lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1005.196734] env[65522]: DEBUG nova.network.neutron [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Refreshing network info cache for port ed6e9224-69f8-46fd-a71f-27d8bc6eaec3 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1005.400083] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114529, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483331} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.400396] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 8859b051-8f75-4aad-b789-42662019d4c5/8859b051-8f75-4aad-b789-42662019d4c5.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1005.400610] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1005.400928] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e8d2d54e-5295-4e52-9da7-c6932ff60c5c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.411070] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1005.411070] env[65522]: value = "task-5114533" [ 1005.411070] env[65522]: _type = "Task" [ 1005.411070] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.418911] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114528, 'name': CloneVM_Task} progress is 95%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.425571] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114533, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.614909] env[65522]: DEBUG nova.compute.utils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1005.616533] env[65522]: DEBUG nova.compute.manager [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1005.616861] env[65522]: DEBUG nova.network.neutron [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1005.617489] env[65522]: WARNING neutronclient.v2_0.client [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1005.617664] env[65522]: WARNING neutronclient.v2_0.client [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1005.618513] env[65522]: WARNING openstack [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.618641] env[65522]: WARNING openstack [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1005.637095] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114532, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.666348] env[65522]: DEBUG nova.policy [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b491939798e4481fb433ffb81f366b25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '068c2387de8c406194d9b1762c7292a7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1005.701940] env[65522]: WARNING neutronclient.v2_0.client [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1005.701940] env[65522]: WARNING openstack [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.701940] env[65522]: WARNING openstack [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1005.794295] env[65522]: WARNING openstack [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.794656] env[65522]: WARNING openstack [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1005.869474] env[65522]: WARNING neutronclient.v2_0.client [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1005.870238] env[65522]: WARNING openstack [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.870492] env[65522]: WARNING openstack [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1005.923309] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114528, 'name': CloneVM_Task, 'duration_secs': 1.672149} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.923309] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Created linked-clone VM from snapshot [ 1005.924114] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2c365d-7bfd-435b-9507-f275fc6b1514 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.929618] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114533, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.195479} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.930206] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1005.930988] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c928868-c71c-4327-b5e9-e9f71a091aa3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.936329] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Uploading image b0c7a88b-7872-412f-9204-61eaa49a2c35 {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1005.959246] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 8859b051-8f75-4aad-b789-42662019d4c5/8859b051-8f75-4aad-b789-42662019d4c5.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1005.962100] env[65522]: DEBUG nova.network.neutron [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Successfully created port: 2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1005.964072] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-657f1618-3a3b-41f2-bdf9-757b719f2d4a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.981308] env[65522]: DEBUG nova.network.neutron [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updated VIF entry in instance network info cache for port ed6e9224-69f8-46fd-a71f-27d8bc6eaec3. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1005.981658] env[65522]: DEBUG nova.network.neutron [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance_info_cache with network_info: [{"id": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "address": "fa:16:3e:c3:12:7d", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped6e9224-69", "ovs_interfaceid": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1005.988851] env[65522]: DEBUG oslo_vmware.rw_handles [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1005.988851] env[65522]: value = "vm-994882" [ 1005.988851] env[65522]: _type = "VirtualMachine" [ 1005.988851] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1005.989134] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-2b3f1d57-33bd-4e5a-904e-832ba145074a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.991949] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1005.991949] env[65522]: value = "task-5114534" [ 1005.991949] env[65522]: _type = "Task" [ 1005.991949] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.998034] env[65522]: DEBUG oslo_vmware.rw_handles [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lease: (returnval){ [ 1005.998034] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52196e13-0fbd-11af-0c24-dd878ae70f12" [ 1005.998034] env[65522]: _type = "HttpNfcLease" [ 1005.998034] env[65522]: } obtained for exporting VM: (result){ [ 1005.998034] env[65522]: value = "vm-994882" [ 1005.998034] env[65522]: _type = "VirtualMachine" [ 1005.998034] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1005.998384] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the lease: (returnval){ [ 1005.998384] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52196e13-0fbd-11af-0c24-dd878ae70f12" [ 1005.998384] env[65522]: _type = "HttpNfcLease" [ 1005.998384] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1006.004683] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114534, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.012145] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1006.012145] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52196e13-0fbd-11af-0c24-dd878ae70f12" [ 1006.012145] env[65522]: _type = "HttpNfcLease" [ 1006.012145] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1006.127282] env[65522]: DEBUG nova.compute.manager [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1006.142085] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114532, 'name': CreateVM_Task, 'duration_secs': 0.535771} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.142085] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1006.142705] env[65522]: WARNING neutronclient.v2_0.client [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1006.143292] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.143574] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1006.144037] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1006.148975] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac7abea7-4bc7-44f4-99d1-d94d0fa87be1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.155652] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1006.155652] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52720e63-c175-5f11-e87f-34c349db69c4" [ 1006.155652] env[65522]: _type = "Task" [ 1006.155652] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.170540] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52720e63-c175-5f11-e87f-34c349db69c4, 'name': SearchDatastore_Task, 'duration_secs': 0.010404} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.170867] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1006.171117] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1006.171367] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.171510] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1006.171689] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1006.172710] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-caae5f00-046a-4e0a-9163-4f56bdb716a7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.182804] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1006.183015] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1006.183789] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e649280-a95b-4c8a-aab3-7211bf16bea8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.192938] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1006.192938] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52369c86-fbde-6ffb-83c2-fa62a584db84" [ 1006.192938] env[65522]: _type = "Task" [ 1006.192938] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.202845] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52369c86-fbde-6ffb-83c2-fa62a584db84, 'name': SearchDatastore_Task, 'duration_secs': 0.009395} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.206561] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56b67ed3-37f3-4718-bd16-d464c33a67b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.212554] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1006.212554] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ba6288-e4b6-6e1e-9f64-9348a88722e2" [ 1006.212554] env[65522]: _type = "Task" [ 1006.212554] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.226597] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ba6288-e4b6-6e1e-9f64-9348a88722e2, 'name': SearchDatastore_Task, 'duration_secs': 0.010838} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.226597] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1006.226881] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 9f342b89-bde2-4c35-ae42-cfe1e6973b74/9f342b89-bde2-4c35-ae42-cfe1e6973b74.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1006.227212] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3db2c2ef-c73a-42e3-bc73-5a1a3ca8a9b5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.235525] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1006.235525] env[65522]: value = "task-5114536" [ 1006.235525] env[65522]: _type = "Task" [ 1006.235525] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.246554] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114536, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.440354] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756d9474-37c7-48c8-bd02-0b7bc765a24b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.450522] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-284a86cc-ba86-4b4d-b5e7-512b583ff5c8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.483884] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad612e9-f74c-4a25-8200-f340deb040d8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.487687] env[65522]: DEBUG oslo_concurrency.lockutils [req-d0b407f8-99fa-4f6d-8620-21365b7b5492 req-cc7bae80-80ad-4194-b046-731116360b13 service nova] Releasing lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1006.498685] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3131774f-8ded-4d70-8fcc-2cae12cf9a40 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.512712] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114534, 'name': ReconfigVM_Task, 'duration_secs': 0.358776} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.521429] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 8859b051-8f75-4aad-b789-42662019d4c5/8859b051-8f75-4aad-b789-42662019d4c5.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.522666] env[65522]: DEBUG nova.compute.provider_tree [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.523907] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-419523e3-5c2c-4c4e-80dc-65e67d5bd0c7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.527647] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1006.527647] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52196e13-0fbd-11af-0c24-dd878ae70f12" [ 1006.527647] env[65522]: _type = "HttpNfcLease" [ 1006.527647] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1006.528396] env[65522]: DEBUG oslo_vmware.rw_handles [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1006.528396] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52196e13-0fbd-11af-0c24-dd878ae70f12" [ 1006.528396] env[65522]: _type = "HttpNfcLease" [ 1006.528396] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1006.529284] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d8efd0-026f-4b81-9b08-df9a67e069c0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.535756] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1006.535756] env[65522]: value = "task-5114537" [ 1006.535756] env[65522]: _type = "Task" [ 1006.535756] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.543653] env[65522]: DEBUG oslo_vmware.rw_handles [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e6a733-8a95-e3df-a94e-c34a55047673/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1006.543877] env[65522]: DEBUG oslo_vmware.rw_handles [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e6a733-8a95-e3df-a94e-c34a55047673/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1006.603544] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114537, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.747581] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114536, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.765176] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f1cc58b5-533b-499c-bc51-5add4fccd799 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.028828] env[65522]: DEBUG nova.scheduler.client.report [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1007.050872] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114537, 'name': Rename_Task, 'duration_secs': 0.336446} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.051978] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1007.052344] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dbb3e890-8651-4cc1-a890-d5d870743a6f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.062119] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1007.062119] env[65522]: value = "task-5114538" [ 1007.062119] env[65522]: _type = "Task" [ 1007.062119] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.071529] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114538, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.142263] env[65522]: DEBUG nova.compute.manager [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1007.167969] env[65522]: DEBUG nova.virt.hardware [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1007.168255] env[65522]: DEBUG nova.virt.hardware [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1007.168410] env[65522]: DEBUG nova.virt.hardware [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1007.168592] env[65522]: DEBUG nova.virt.hardware [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1007.168862] env[65522]: DEBUG nova.virt.hardware [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1007.169072] env[65522]: DEBUG nova.virt.hardware [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1007.169292] env[65522]: DEBUG nova.virt.hardware [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1007.169422] env[65522]: DEBUG nova.virt.hardware [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1007.169555] env[65522]: DEBUG nova.virt.hardware [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1007.169713] env[65522]: DEBUG nova.virt.hardware [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1007.169883] env[65522]: DEBUG nova.virt.hardware [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1007.170860] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694e15d0-78d3-4af6-bf09-d4ab3c377cba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.179835] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd2b59e-3d8a-44ef-8caa-a4802e9ee074 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.246174] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114536, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539365} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.246583] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 9f342b89-bde2-4c35-ae42-cfe1e6973b74/9f342b89-bde2-4c35-ae42-cfe1e6973b74.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1007.247010] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1007.247386] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-60bcb38b-175a-46b8-b694-34fa7234bb62 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.256316] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1007.256316] env[65522]: value = "task-5114539" [ 1007.256316] env[65522]: _type = "Task" [ 1007.256316] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.267662] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114539, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.482166] env[65522]: DEBUG nova.compute.manager [req-2afb3487-d4d0-4fdc-9360-c3558eea891b req-6728ac8b-8a28-4c4e-800f-a12abf3ca729 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Received event network-vif-plugged-2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1007.482621] env[65522]: DEBUG oslo_concurrency.lockutils [req-2afb3487-d4d0-4fdc-9360-c3558eea891b req-6728ac8b-8a28-4c4e-800f-a12abf3ca729 service nova] Acquiring lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1007.482950] env[65522]: DEBUG oslo_concurrency.lockutils [req-2afb3487-d4d0-4fdc-9360-c3558eea891b req-6728ac8b-8a28-4c4e-800f-a12abf3ca729 service nova] Lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1007.483130] env[65522]: DEBUG oslo_concurrency.lockutils [req-2afb3487-d4d0-4fdc-9360-c3558eea891b req-6728ac8b-8a28-4c4e-800f-a12abf3ca729 service nova] Lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1007.483389] env[65522]: DEBUG nova.compute.manager [req-2afb3487-d4d0-4fdc-9360-c3558eea891b req-6728ac8b-8a28-4c4e-800f-a12abf3ca729 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] No waiting events found dispatching network-vif-plugged-2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1007.484384] env[65522]: WARNING nova.compute.manager [req-2afb3487-d4d0-4fdc-9360-c3558eea891b req-6728ac8b-8a28-4c4e-800f-a12abf3ca729 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Received unexpected event network-vif-plugged-2b2de45d-4c08-4bb4-810a-2f26266cbc20 for instance with vm_state building and task_state spawning. [ 1007.536982] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.431s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1007.536982] env[65522]: DEBUG nova.compute.manager [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1007.540299] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.226s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1007.542307] env[65522]: INFO nova.compute.claims [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1007.574813] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114538, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.587693] env[65522]: DEBUG nova.network.neutron [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Successfully updated port: 2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1007.766960] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114539, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.116818} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.767245] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1007.768346] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74aacef-467b-4739-b556-94bc3065f82b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.793921] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 9f342b89-bde2-4c35-ae42-cfe1e6973b74/9f342b89-bde2-4c35-ae42-cfe1e6973b74.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1007.794748] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35f398a3-e2c5-4865-a486-63b7b497c620 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.818083] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1007.818083] env[65522]: value = "task-5114540" [ 1007.818083] env[65522]: _type = "Task" [ 1007.818083] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.828695] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114540, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.050685] env[65522]: DEBUG nova.compute.utils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1008.052842] env[65522]: DEBUG nova.compute.manager [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1008.053080] env[65522]: DEBUG nova.network.neutron [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1008.053565] env[65522]: WARNING neutronclient.v2_0.client [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1008.053869] env[65522]: WARNING neutronclient.v2_0.client [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1008.054457] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1008.054806] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.073854] env[65522]: DEBUG oslo_vmware.api [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114538, 'name': PowerOnVM_Task, 'duration_secs': 0.940363} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.075859] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1008.075859] env[65522]: INFO nova.compute.manager [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Took 7.69 seconds to spawn the instance on the hypervisor. [ 1008.075859] env[65522]: DEBUG nova.compute.manager [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1008.076952] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6471e8-97d7-4b3d-ae4c-7c2879bb87b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.094041] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.094041] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1008.094041] env[65522]: DEBUG nova.network.neutron [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1008.130339] env[65522]: DEBUG nova.policy [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b7e48c7c3604176b4f844bcd255cb01', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9db9e23ad1b745afa496de2deae83865', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1008.330137] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114540, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.503939] env[65522]: DEBUG nova.network.neutron [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Successfully created port: 44e3c08b-10b8-44fb-9e3e-1d2f77752195 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1008.553778] env[65522]: DEBUG nova.compute.manager [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1008.602686] env[65522]: WARNING openstack [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1008.603244] env[65522]: WARNING openstack [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.614456] env[65522]: INFO nova.compute.manager [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Took 28.51 seconds to build instance. [ 1008.651378] env[65522]: DEBUG nova.network.neutron [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1008.674470] env[65522]: WARNING openstack [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1008.675123] env[65522]: WARNING openstack [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.769687] env[65522]: WARNING neutronclient.v2_0.client [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1008.770410] env[65522]: WARNING openstack [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1008.770796] env[65522]: WARNING openstack [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.835787] env[65522]: DEBUG nova.network.neutron [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Successfully created port: 417e342f-fd78-422e-8ae2-05691d9ff64c {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1008.840650] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114540, 'name': ReconfigVM_Task, 'duration_secs': 0.532476} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.840969] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 9f342b89-bde2-4c35-ae42-cfe1e6973b74/9f342b89-bde2-4c35-ae42-cfe1e6973b74.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1008.841697] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-884207c4-abcf-409a-9c8e-9929cc3bece8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.851948] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1008.851948] env[65522]: value = "task-5114541" [ 1008.851948] env[65522]: _type = "Task" [ 1008.851948] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.862914] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114541, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.878109] env[65522]: DEBUG nova.network.neutron [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updating instance_info_cache with network_info: [{"id": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "address": "fa:16:3e:e9:bd:f6", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b2de45d-4c", "ovs_interfaceid": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1008.933739] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6824c204-6258-4c8a-91cc-0494ea7fad39 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.942725] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e99cccb0-e8b7-466e-bc34-b6b192e5c9b4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.977252] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824dde70-ea8c-4118-8d7f-99c2554ef161 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.986605] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ac5fd7-fd01-4f1c-bb52-aeef51c70c68 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.002477] env[65522]: DEBUG nova.compute.provider_tree [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.119047] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91193899-0e67-41b4-ba76-c46237632358 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "8859b051-8f75-4aad-b789-42662019d4c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.025s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1009.125386] env[65522]: DEBUG nova.network.neutron [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Successfully created port: 7052a26d-df5e-4222-be00-5bf53feb656f {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1009.366246] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114541, 'name': Rename_Task, 'duration_secs': 0.278665} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.367042] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1009.367042] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-592ecc3e-0e73-4a5f-a5ec-892c7a23609d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.376099] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1009.376099] env[65522]: value = "task-5114542" [ 1009.376099] env[65522]: _type = "Task" [ 1009.376099] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.393125] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1009.393531] env[65522]: DEBUG nova.compute.manager [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Instance network_info: |[{"id": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "address": "fa:16:3e:e9:bd:f6", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b2de45d-4c", "ovs_interfaceid": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1009.393894] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114542, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.394400] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e9:bd:f6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4345ef6-a7c8-4c1c-badf-a0d4f578b61c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b2de45d-4c08-4bb4-810a-2f26266cbc20', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1009.402132] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1009.402389] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1009.402633] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b2e19ee7-aa1d-47f2-b094-5d840f16eb12 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.425498] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.425498] env[65522]: value = "task-5114543" [ 1009.425498] env[65522]: _type = "Task" [ 1009.425498] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.435220] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114543, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.509075] env[65522]: DEBUG nova.scheduler.client.report [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1009.568971] env[65522]: DEBUG nova.compute.manager [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1009.600554] env[65522]: DEBUG nova.virt.hardware [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1009.600769] env[65522]: DEBUG nova.virt.hardware [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1009.601177] env[65522]: DEBUG nova.virt.hardware [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1009.601177] env[65522]: DEBUG nova.virt.hardware [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1009.601308] env[65522]: DEBUG nova.virt.hardware [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1009.601446] env[65522]: DEBUG nova.virt.hardware [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1009.601727] env[65522]: DEBUG nova.virt.hardware [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1009.601937] env[65522]: DEBUG nova.virt.hardware [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1009.602167] env[65522]: DEBUG nova.virt.hardware [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1009.602299] env[65522]: DEBUG nova.virt.hardware [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1009.602474] env[65522]: DEBUG nova.virt.hardware [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1009.603703] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc424fd3-34f2-4f7d-8e9c-7a1a95293703 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.614876] env[65522]: DEBUG nova.compute.manager [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Received event network-changed-2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1009.614985] env[65522]: DEBUG nova.compute.manager [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Refreshing instance network info cache due to event network-changed-2b2de45d-4c08-4bb4-810a-2f26266cbc20. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1009.615636] env[65522]: DEBUG oslo_concurrency.lockutils [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] Acquiring lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.615636] env[65522]: DEBUG oslo_concurrency.lockutils [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] Acquired lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1009.615636] env[65522]: DEBUG nova.network.neutron [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Refreshing network info cache for port 2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1009.617649] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22c62c0-367b-4caf-89f8-54097452f02c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.885686] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114542, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.935757] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114543, 'name': CreateVM_Task, 'duration_secs': 0.447005} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.935931] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1009.936455] env[65522]: WARNING neutronclient.v2_0.client [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1009.936817] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.936998] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1009.937325] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1009.937595] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c3f9a9c-d062-413e-85f5-2461660cb09c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.943051] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1009.943051] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d70cdf-da67-bd26-c01b-2e89d8f57db5" [ 1009.943051] env[65522]: _type = "Task" [ 1009.943051] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.951920] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d70cdf-da67-bd26-c01b-2e89d8f57db5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.013120] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1010.013668] env[65522]: DEBUG nova.compute.manager [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1010.016398] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.197s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1010.016590] env[65522]: DEBUG nova.objects.instance [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1010.124351] env[65522]: WARNING neutronclient.v2_0.client [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1010.124637] env[65522]: WARNING openstack [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1010.124996] env[65522]: WARNING openstack [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1010.386778] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114542, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.405667] env[65522]: WARNING openstack [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1010.406067] env[65522]: WARNING openstack [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1010.460431] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d70cdf-da67-bd26-c01b-2e89d8f57db5, 'name': SearchDatastore_Task, 'duration_secs': 0.025858} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.460738] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1010.461020] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.461275] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.461406] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1010.461575] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.461874] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a13d4f7a-1d37-45a6-9b52-e38eb0dbcf83 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.472957] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.473181] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1010.473924] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea788907-facc-43c2-a452-b35d0cc1fa82 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.480485] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1010.480485] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52017442-0563-b995-a498-e11a04cb7499" [ 1010.480485] env[65522]: _type = "Task" [ 1010.480485] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.490226] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52017442-0563-b995-a498-e11a04cb7499, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.520643] env[65522]: WARNING neutronclient.v2_0.client [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1010.521452] env[65522]: WARNING openstack [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1010.521845] env[65522]: WARNING openstack [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1010.531817] env[65522]: DEBUG nova.compute.utils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1010.537524] env[65522]: DEBUG nova.compute.manager [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1010.537524] env[65522]: DEBUG nova.network.neutron [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1010.537988] env[65522]: WARNING neutronclient.v2_0.client [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1010.538373] env[65522]: WARNING neutronclient.v2_0.client [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1010.539153] env[65522]: WARNING openstack [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1010.539458] env[65522]: WARNING openstack [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1010.600484] env[65522]: DEBUG nova.compute.manager [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1010.601604] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e864fcd9-6a12-496b-9548-8a1388cb4396 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.641105] env[65522]: DEBUG nova.compute.manager [req-8deb4737-b666-475e-b68f-25ae2e7412b1 req-4365e54f-ddc8-4490-aba8-14eee66575f0 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Received event network-vif-plugged-44e3c08b-10b8-44fb-9e3e-1d2f77752195 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1010.641686] env[65522]: DEBUG oslo_concurrency.lockutils [req-8deb4737-b666-475e-b68f-25ae2e7412b1 req-4365e54f-ddc8-4490-aba8-14eee66575f0 service nova] Acquiring lock "e7baf2a0-21dd-4610-8230-81b99b64856b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1010.641994] env[65522]: DEBUG oslo_concurrency.lockutils [req-8deb4737-b666-475e-b68f-25ae2e7412b1 req-4365e54f-ddc8-4490-aba8-14eee66575f0 service nova] Lock "e7baf2a0-21dd-4610-8230-81b99b64856b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1010.642167] env[65522]: DEBUG oslo_concurrency.lockutils [req-8deb4737-b666-475e-b68f-25ae2e7412b1 req-4365e54f-ddc8-4490-aba8-14eee66575f0 service nova] Lock "e7baf2a0-21dd-4610-8230-81b99b64856b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1010.642332] env[65522]: DEBUG nova.compute.manager [req-8deb4737-b666-475e-b68f-25ae2e7412b1 req-4365e54f-ddc8-4490-aba8-14eee66575f0 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] No waiting events found dispatching network-vif-plugged-44e3c08b-10b8-44fb-9e3e-1d2f77752195 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1010.642448] env[65522]: WARNING nova.compute.manager [req-8deb4737-b666-475e-b68f-25ae2e7412b1 req-4365e54f-ddc8-4490-aba8-14eee66575f0 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Received unexpected event network-vif-plugged-44e3c08b-10b8-44fb-9e3e-1d2f77752195 for instance with vm_state building and task_state spawning. [ 1010.651905] env[65522]: DEBUG nova.policy [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e74c132f591d4c6fbe65a75c1e1e4a24', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ecccb656b0d4c96b40b200cdcddbad5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1010.706897] env[65522]: DEBUG nova.network.neutron [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updated VIF entry in instance network info cache for port 2b2de45d-4c08-4bb4-810a-2f26266cbc20. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1010.707333] env[65522]: DEBUG nova.network.neutron [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updating instance_info_cache with network_info: [{"id": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "address": "fa:16:3e:e9:bd:f6", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b2de45d-4c", "ovs_interfaceid": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1010.775302] env[65522]: DEBUG nova.network.neutron [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Successfully updated port: 44e3c08b-10b8-44fb-9e3e-1d2f77752195 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1010.886715] env[65522]: DEBUG oslo_vmware.api [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114542, 'name': PowerOnVM_Task, 'duration_secs': 1.258475} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.887080] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1010.887269] env[65522]: INFO nova.compute.manager [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Took 8.05 seconds to spawn the instance on the hypervisor. [ 1010.887366] env[65522]: DEBUG nova.compute.manager [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1010.888188] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b2a150-9b75-47c2-96a0-4cfac35f1568 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.001462] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52017442-0563-b995-a498-e11a04cb7499, 'name': SearchDatastore_Task, 'duration_secs': 0.018559} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.001462] env[65522]: DEBUG nova.network.neutron [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Successfully created port: 15e24bea-7053-40cb-8a80-6782dcca755f {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1011.006227] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56b8f68b-2e33-4c57-ad9d-b6afe26d8fc6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.011301] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1011.011301] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525c75ee-6039-408f-7b31-e148cb423df4" [ 1011.011301] env[65522]: _type = "Task" [ 1011.011301] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.020455] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525c75ee-6039-408f-7b31-e148cb423df4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.038028] env[65522]: DEBUG nova.compute.manager [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1011.051262] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d420c6ed-1ba4-4075-a433-d9d55b05b9f2 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.032s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1011.051262] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.541s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1011.051262] env[65522]: DEBUG nova.objects.instance [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lazy-loading 'resources' on Instance uuid 87a3c63d-794d-44ab-bad6-65c323d72ae7 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.119600] env[65522]: INFO nova.compute.manager [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] instance snapshotting [ 1011.122569] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02410e0d-fe5f-4561-9947-ac0fee80682b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.142496] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27f3258-3f18-42fb-944f-227319c5ad3b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.210546] env[65522]: DEBUG oslo_concurrency.lockutils [req-683ce1b1-7026-4763-a125-d80fce70ec89 req-6faeb36f-b0be-441d-83c0-4a304f743e64 service nova] Releasing lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1011.408456] env[65522]: INFO nova.compute.manager [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Took 30.79 seconds to build instance. [ 1011.522631] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525c75ee-6039-408f-7b31-e148cb423df4, 'name': SearchDatastore_Task, 'duration_secs': 0.015686} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.523051] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1011.523304] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3/11ef190f-8ade-4705-b3b0-a8ff8b97bcd3.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1011.523676] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a76c24af-d8c7-4678-b456-2597d5e27cd6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.535026] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1011.535026] env[65522]: value = "task-5114544" [ 1011.535026] env[65522]: _type = "Task" [ 1011.535026] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.547575] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114544, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.654881] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1011.655472] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cc3d7794-ba19-4416-ad5a-531aec6c7b08 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.664068] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1011.664068] env[65522]: value = "task-5114545" [ 1011.664068] env[65522]: _type = "Task" [ 1011.664068] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.676679] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114545, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.875487] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8bec36-bb2e-49fb-9e5b-740036c72ce5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.884738] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0f8f71-a823-4d63-aec5-a08da19b5c76 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.918331] env[65522]: DEBUG oslo_concurrency.lockutils [None req-26a52760-8a96-4acc-9cb7-e8d4602dc702 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.316s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1011.919547] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1522fb7-cc5e-4170-95c7-a5360844c1aa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.928641] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215476a6-59b4-477d-81fe-71ee4b292890 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.944649] env[65522]: DEBUG nova.compute.provider_tree [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.042806] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114544, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.049129] env[65522]: DEBUG nova.compute.manager [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1012.078162] env[65522]: DEBUG nova.virt.hardware [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1012.078435] env[65522]: DEBUG nova.virt.hardware [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1012.078595] env[65522]: DEBUG nova.virt.hardware [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1012.078809] env[65522]: DEBUG nova.virt.hardware [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1012.078978] env[65522]: DEBUG nova.virt.hardware [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1012.079148] env[65522]: DEBUG nova.virt.hardware [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1012.079358] env[65522]: DEBUG nova.virt.hardware [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.079515] env[65522]: DEBUG nova.virt.hardware [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1012.079679] env[65522]: DEBUG nova.virt.hardware [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1012.079841] env[65522]: DEBUG nova.virt.hardware [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1012.080060] env[65522]: DEBUG nova.virt.hardware [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1012.081065] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc91598a-41fa-4ff6-8ce6-aca953253c0d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.094181] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a9611de-7400-4998-9f2b-d7a788d1840a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.176188] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114545, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.448383] env[65522]: DEBUG nova.scheduler.client.report [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1012.542274] env[65522]: DEBUG nova.network.neutron [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Successfully updated port: 15e24bea-7053-40cb-8a80-6782dcca755f {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1012.547607] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114544, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.87893} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.548176] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3/11ef190f-8ade-4705-b3b0-a8ff8b97bcd3.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1012.548403] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1012.548902] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c0fb7ea-b769-4838-aa50-6625ab9b43e4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.557301] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1012.557301] env[65522]: value = "task-5114546" [ 1012.557301] env[65522]: _type = "Task" [ 1012.557301] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.569311] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114546, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.677073] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114545, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.863582] env[65522]: DEBUG nova.network.neutron [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Successfully updated port: 417e342f-fd78-422e-8ae2-05691d9ff64c {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1012.956319] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.906s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1012.961212] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.796s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1012.961212] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1012.961809] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.984s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1012.966094] env[65522]: INFO nova.compute.claims [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1013.031931] env[65522]: INFO nova.scheduler.client.report [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleted allocations for instance a2a5fb2d-62e0-4809-a01f-f1df66dca58f [ 1013.035676] env[65522]: INFO nova.scheduler.client.report [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Deleted allocations for instance 87a3c63d-794d-44ab-bad6-65c323d72ae7 [ 1013.049879] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "refresh_cache-5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.050117] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "refresh_cache-5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1013.050315] env[65522]: DEBUG nova.network.neutron [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1013.065184] env[65522]: DEBUG nova.compute.manager [req-9e740ad3-fcfc-4a61-8f97-c97f6ea22de3 req-6b13ef2f-2ed8-4c14-a1be-97113bd8e96f service nova] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Received event network-vif-plugged-15e24bea-7053-40cb-8a80-6782dcca755f {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1013.065447] env[65522]: DEBUG oslo_concurrency.lockutils [req-9e740ad3-fcfc-4a61-8f97-c97f6ea22de3 req-6b13ef2f-2ed8-4c14-a1be-97113bd8e96f service nova] Acquiring lock "5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.065763] env[65522]: DEBUG oslo_concurrency.lockutils [req-9e740ad3-fcfc-4a61-8f97-c97f6ea22de3 req-6b13ef2f-2ed8-4c14-a1be-97113bd8e96f service nova] Lock "5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.065867] env[65522]: DEBUG oslo_concurrency.lockutils [req-9e740ad3-fcfc-4a61-8f97-c97f6ea22de3 req-6b13ef2f-2ed8-4c14-a1be-97113bd8e96f service nova] Lock "5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.066669] env[65522]: DEBUG nova.compute.manager [req-9e740ad3-fcfc-4a61-8f97-c97f6ea22de3 req-6b13ef2f-2ed8-4c14-a1be-97113bd8e96f service nova] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] No waiting events found dispatching network-vif-plugged-15e24bea-7053-40cb-8a80-6782dcca755f {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1013.066669] env[65522]: WARNING nova.compute.manager [req-9e740ad3-fcfc-4a61-8f97-c97f6ea22de3 req-6b13ef2f-2ed8-4c14-a1be-97113bd8e96f service nova] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Received unexpected event network-vif-plugged-15e24bea-7053-40cb-8a80-6782dcca755f for instance with vm_state building and task_state spawning. [ 1013.087827] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114546, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080096} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.088758] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1013.089894] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015f0bf1-8dce-44c2-a761-13b2e9618936 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.119078] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3/11ef190f-8ade-4705-b3b0-a8ff8b97bcd3.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.120429] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2173085e-1f83-4ff0-a47f-722f67d11664 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.138837] env[65522]: DEBUG nova.compute.manager [req-97840ffc-9260-4201-a1e0-9d59adb6b94c req-81c0721d-aac3-4771-9dd6-109161a24ac3 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Received event network-changed-44e3c08b-10b8-44fb-9e3e-1d2f77752195 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1013.139070] env[65522]: DEBUG nova.compute.manager [req-97840ffc-9260-4201-a1e0-9d59adb6b94c req-81c0721d-aac3-4771-9dd6-109161a24ac3 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Refreshing instance network info cache due to event network-changed-44e3c08b-10b8-44fb-9e3e-1d2f77752195. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1013.139296] env[65522]: DEBUG oslo_concurrency.lockutils [req-97840ffc-9260-4201-a1e0-9d59adb6b94c req-81c0721d-aac3-4771-9dd6-109161a24ac3 service nova] Acquiring lock "refresh_cache-e7baf2a0-21dd-4610-8230-81b99b64856b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.139433] env[65522]: DEBUG oslo_concurrency.lockutils [req-97840ffc-9260-4201-a1e0-9d59adb6b94c req-81c0721d-aac3-4771-9dd6-109161a24ac3 service nova] Acquired lock "refresh_cache-e7baf2a0-21dd-4610-8230-81b99b64856b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1013.139588] env[65522]: DEBUG nova.network.neutron [req-97840ffc-9260-4201-a1e0-9d59adb6b94c req-81c0721d-aac3-4771-9dd6-109161a24ac3 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Refreshing network info cache for port 44e3c08b-10b8-44fb-9e3e-1d2f77752195 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1013.149501] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1013.149501] env[65522]: value = "task-5114547" [ 1013.149501] env[65522]: _type = "Task" [ 1013.149501] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.160397] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114547, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.179671] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114545, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.548728] env[65522]: DEBUG oslo_concurrency.lockutils [None req-84af248f-ce64-4bd8-9176-5281dc3443c2 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "87a3c63d-794d-44ab-bad6-65c323d72ae7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.004s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.553091] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4976af99-11d4-4d14-bb1b-e49d0675fcdd tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "a2a5fb2d-62e0-4809-a01f-f1df66dca58f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.866s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.557243] env[65522]: WARNING openstack [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1013.559630] env[65522]: WARNING openstack [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1013.643416] env[65522]: WARNING neutronclient.v2_0.client [req-97840ffc-9260-4201-a1e0-9d59adb6b94c req-81c0721d-aac3-4771-9dd6-109161a24ac3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1013.645423] env[65522]: WARNING openstack [req-97840ffc-9260-4201-a1e0-9d59adb6b94c req-81c0721d-aac3-4771-9dd6-109161a24ac3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1013.645423] env[65522]: WARNING openstack [req-97840ffc-9260-4201-a1e0-9d59adb6b94c req-81c0721d-aac3-4771-9dd6-109161a24ac3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1013.662780] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114547, 'name': ReconfigVM_Task, 'duration_secs': 0.32848} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.664121] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3/11ef190f-8ade-4705-b3b0-a8ff8b97bcd3.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.664273] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-04635f66-4426-458a-8f66-873a7e626819 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.673092] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1013.673092] env[65522]: value = "task-5114548" [ 1013.673092] env[65522]: _type = "Task" [ 1013.673092] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.680208] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114545, 'name': CreateSnapshot_Task, 'duration_secs': 1.674207} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.680903] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1013.681747] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30580145-a48c-4d2d-9760-7665ec2cdf3c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.688026] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114548, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.761110] env[65522]: DEBUG nova.network.neutron [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1013.827190] env[65522]: DEBUG nova.network.neutron [req-97840ffc-9260-4201-a1e0-9d59adb6b94c req-81c0721d-aac3-4771-9dd6-109161a24ac3 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1013.841576] env[65522]: WARNING openstack [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1013.842141] env[65522]: WARNING openstack [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.113032] env[65522]: DEBUG nova.network.neutron [req-97840ffc-9260-4201-a1e0-9d59adb6b94c req-81c0721d-aac3-4771-9dd6-109161a24ac3 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1014.146466] env[65522]: WARNING neutronclient.v2_0.client [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.147158] env[65522]: WARNING openstack [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.147522] env[65522]: WARNING openstack [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.187493] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114548, 'name': Rename_Task, 'duration_secs': 0.165159} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.190486] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1014.193401] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0c94179b-19b4-478e-a16f-6bfd2d7d9a39 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.204880] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1014.206830] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6ce6b223-2795-4040-ae4a-83da0daab857 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.210948] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1014.210948] env[65522]: value = "task-5114549" [ 1014.210948] env[65522]: _type = "Task" [ 1014.210948] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.220913] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1014.220913] env[65522]: value = "task-5114550" [ 1014.220913] env[65522]: _type = "Task" [ 1014.220913] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.230319] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114549, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.246529] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114550, 'name': CloneVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.303175] env[65522]: DEBUG nova.network.neutron [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Updating instance_info_cache with network_info: [{"id": "15e24bea-7053-40cb-8a80-6782dcca755f", "address": "fa:16:3e:cc:52:4b", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15e24bea-70", "ovs_interfaceid": "15e24bea-7053-40cb-8a80-6782dcca755f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1014.327231] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4ec8d7-2438-45a3-b8b3-92eb2035e1ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.338680] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b01632-3d48-48cc-8be9-2101fd0c8ad5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.380178] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8188022f-1808-4978-b365-84dc543ac896 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.394712] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85534ad9-32c6-4ff7-9994-8f755a021c94 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.412623] env[65522]: DEBUG nova.compute.provider_tree [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.618800] env[65522]: DEBUG oslo_concurrency.lockutils [req-97840ffc-9260-4201-a1e0-9d59adb6b94c req-81c0721d-aac3-4771-9dd6-109161a24ac3 service nova] Releasing lock "refresh_cache-e7baf2a0-21dd-4610-8230-81b99b64856b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1014.726725] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114549, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.741649] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114550, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.807222] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "refresh_cache-5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1014.807795] env[65522]: DEBUG nova.compute.manager [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Instance network_info: |[{"id": "15e24bea-7053-40cb-8a80-6782dcca755f", "address": "fa:16:3e:cc:52:4b", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15e24bea-70", "ovs_interfaceid": "15e24bea-7053-40cb-8a80-6782dcca755f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1014.808456] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:52:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6fb0104-186b-4288-b87e-634893f46f01', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '15e24bea-7053-40cb-8a80-6782dcca755f', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1014.820312] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1014.820679] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1014.821280] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c1919fe-e960-4762-ab8f-8c0018fecae1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.844483] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1014.844483] env[65522]: value = "task-5114551" [ 1014.844483] env[65522]: _type = "Task" [ 1014.844483] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.854141] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114551, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.915984] env[65522]: DEBUG nova.scheduler.client.report [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1014.995188] env[65522]: DEBUG nova.network.neutron [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Successfully updated port: 7052a26d-df5e-4222-be00-5bf53feb656f {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1015.097441] env[65522]: DEBUG nova.compute.manager [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Received event network-changed-15e24bea-7053-40cb-8a80-6782dcca755f {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1015.097634] env[65522]: DEBUG nova.compute.manager [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Refreshing instance network info cache due to event network-changed-15e24bea-7053-40cb-8a80-6782dcca755f. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1015.097854] env[65522]: DEBUG oslo_concurrency.lockutils [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Acquiring lock "refresh_cache-5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.097998] env[65522]: DEBUG oslo_concurrency.lockutils [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Acquired lock "refresh_cache-5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1015.098160] env[65522]: DEBUG nova.network.neutron [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Refreshing network info cache for port 15e24bea-7053-40cb-8a80-6782dcca755f {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1015.107748] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "ca964440-5375-4aff-8b45-96fbe829dd16" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1015.107983] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "ca964440-5375-4aff-8b45-96fbe829dd16" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1015.109038] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "ca964440-5375-4aff-8b45-96fbe829dd16-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1015.109038] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "ca964440-5375-4aff-8b45-96fbe829dd16-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1015.109038] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "ca964440-5375-4aff-8b45-96fbe829dd16-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1015.110704] env[65522]: INFO nova.compute.manager [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Terminating instance [ 1015.223124] env[65522]: DEBUG nova.compute.manager [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Received event network-vif-plugged-417e342f-fd78-422e-8ae2-05691d9ff64c {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1015.223384] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Acquiring lock "e7baf2a0-21dd-4610-8230-81b99b64856b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1015.223663] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Lock "e7baf2a0-21dd-4610-8230-81b99b64856b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1015.223930] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Lock "e7baf2a0-21dd-4610-8230-81b99b64856b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1015.224081] env[65522]: DEBUG nova.compute.manager [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] No waiting events found dispatching network-vif-plugged-417e342f-fd78-422e-8ae2-05691d9ff64c {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1015.224203] env[65522]: WARNING nova.compute.manager [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Received unexpected event network-vif-plugged-417e342f-fd78-422e-8ae2-05691d9ff64c for instance with vm_state building and task_state spawning. [ 1015.224385] env[65522]: DEBUG nova.compute.manager [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Received event network-changed-417e342f-fd78-422e-8ae2-05691d9ff64c {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1015.224546] env[65522]: DEBUG nova.compute.manager [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Refreshing instance network info cache due to event network-changed-417e342f-fd78-422e-8ae2-05691d9ff64c. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1015.224734] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Acquiring lock "refresh_cache-e7baf2a0-21dd-4610-8230-81b99b64856b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.224910] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Acquired lock "refresh_cache-e7baf2a0-21dd-4610-8230-81b99b64856b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1015.225115] env[65522]: DEBUG nova.network.neutron [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Refreshing network info cache for port 417e342f-fd78-422e-8ae2-05691d9ff64c {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1015.233523] env[65522]: DEBUG oslo_vmware.api [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114549, 'name': PowerOnVM_Task, 'duration_secs': 0.538635} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.237549] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1015.239176] env[65522]: INFO nova.compute.manager [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Took 8.10 seconds to spawn the instance on the hypervisor. [ 1015.239176] env[65522]: DEBUG nova.compute.manager [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1015.239910] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6925c9ed-40f0-4edf-ac5e-51043d29bdde {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.252033] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114550, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.359058] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114551, 'name': CreateVM_Task, 'duration_secs': 0.483183} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.359058] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1015.359221] env[65522]: WARNING neutronclient.v2_0.client [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1015.359520] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.359675] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1015.359978] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1015.360255] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc09b01e-8139-4635-a00c-2b054f500cc8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.368210] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 1015.368210] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5293c87f-a937-3a07-2092-09d4b20665c2" [ 1015.368210] env[65522]: _type = "Task" [ 1015.368210] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.380896] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5293c87f-a937-3a07-2092-09d4b20665c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.422263] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1015.422917] env[65522]: DEBUG nova.compute.manager [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1015.425892] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.602s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1015.426109] env[65522]: DEBUG nova.objects.instance [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1015.497813] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "refresh_cache-e7baf2a0-21dd-4610-8230-81b99b64856b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.601541] env[65522]: WARNING neutronclient.v2_0.client [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1015.602337] env[65522]: WARNING openstack [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.602732] env[65522]: WARNING openstack [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1015.614345] env[65522]: DEBUG nova.compute.manager [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1015.614688] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1015.615886] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23574573-c7f0-4b31-8e68-1ad02a17c1f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.624868] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1015.625281] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-585264f5-ad84-40e1-9137-b5685c569d4f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.636674] env[65522]: DEBUG oslo_vmware.api [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 1015.636674] env[65522]: value = "task-5114552" [ 1015.636674] env[65522]: _type = "Task" [ 1015.636674] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.646983] env[65522]: DEBUG oslo_vmware.api [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114552, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.720885] env[65522]: WARNING openstack [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.721386] env[65522]: WARNING openstack [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1015.732040] env[65522]: WARNING neutronclient.v2_0.client [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1015.732905] env[65522]: WARNING openstack [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.733479] env[65522]: WARNING openstack [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1015.755961] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114550, 'name': CloneVM_Task} progress is 95%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.766985] env[65522]: INFO nova.compute.manager [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Took 30.95 seconds to build instance. [ 1015.792702] env[65522]: DEBUG nova.network.neutron [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1015.813545] env[65522]: WARNING neutronclient.v2_0.client [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1015.814424] env[65522]: WARNING openstack [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.814776] env[65522]: WARNING openstack [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1015.882981] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5293c87f-a937-3a07-2092-09d4b20665c2, 'name': SearchDatastore_Task, 'duration_secs': 0.021889} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.883704] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1015.884160] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1015.884476] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.884631] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1015.884927] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1015.885291] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a84167ae-471d-441b-8a1e-31e909c77b17 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.896477] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1015.896775] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1015.900022] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5564c299-2712-49dc-a278-914f62f33039 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.906195] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 1015.906195] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527cc913-1c9e-3532-0db6-2cf6a05a4a2e" [ 1015.906195] env[65522]: _type = "Task" [ 1015.906195] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.925477] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527cc913-1c9e-3532-0db6-2cf6a05a4a2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.936695] env[65522]: DEBUG nova.compute.utils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1015.944783] env[65522]: DEBUG nova.network.neutron [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1015.946886] env[65522]: DEBUG nova.compute.manager [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1015.947245] env[65522]: DEBUG nova.network.neutron [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1015.947832] env[65522]: WARNING neutronclient.v2_0.client [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1015.948356] env[65522]: WARNING neutronclient.v2_0.client [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1015.949303] env[65522]: WARNING openstack [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.949842] env[65522]: WARNING openstack [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1015.964360] env[65522]: DEBUG nova.compute.manager [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1015.968219] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Releasing lock "refresh_cache-e7baf2a0-21dd-4610-8230-81b99b64856b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1015.968477] env[65522]: DEBUG nova.compute.manager [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Received event network-vif-plugged-7052a26d-df5e-4222-be00-5bf53feb656f {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1015.968781] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Acquiring lock "e7baf2a0-21dd-4610-8230-81b99b64856b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1015.968962] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Lock "e7baf2a0-21dd-4610-8230-81b99b64856b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1015.969191] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Lock "e7baf2a0-21dd-4610-8230-81b99b64856b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1015.969388] env[65522]: DEBUG nova.compute.manager [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] No waiting events found dispatching network-vif-plugged-7052a26d-df5e-4222-be00-5bf53feb656f {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1015.969627] env[65522]: WARNING nova.compute.manager [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Received unexpected event network-vif-plugged-7052a26d-df5e-4222-be00-5bf53feb656f for instance with vm_state building and task_state spawning. [ 1015.969893] env[65522]: DEBUG nova.compute.manager [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Received event network-changed-7052a26d-df5e-4222-be00-5bf53feb656f {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1015.970148] env[65522]: DEBUG nova.compute.manager [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Refreshing instance network info cache due to event network-changed-7052a26d-df5e-4222-be00-5bf53feb656f. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1015.970365] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Acquiring lock "refresh_cache-e7baf2a0-21dd-4610-8230-81b99b64856b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.970545] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Acquired lock "refresh_cache-e7baf2a0-21dd-4610-8230-81b99b64856b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1015.970714] env[65522]: DEBUG nova.network.neutron [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Refreshing network info cache for port 7052a26d-df5e-4222-be00-5bf53feb656f {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1015.974688] env[65522]: DEBUG nova.network.neutron [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Updated VIF entry in instance network info cache for port 15e24bea-7053-40cb-8a80-6782dcca755f. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1015.975249] env[65522]: DEBUG nova.network.neutron [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Updating instance_info_cache with network_info: [{"id": "15e24bea-7053-40cb-8a80-6782dcca755f", "address": "fa:16:3e:cc:52:4b", "network": {"id": "153460fe-78b8-4e8a-935c-806da6533217", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1505394747-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6ecccb656b0d4c96b40b200cdcddbad5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap15e24bea-70", "ovs_interfaceid": "15e24bea-7053-40cb-8a80-6782dcca755f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1016.030035] env[65522]: DEBUG nova.policy [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a97b2f232f7844c993dccdc386ce3b2c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9614f18e01594bf083e619c11760fe36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1016.150522] env[65522]: DEBUG oslo_vmware.api [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114552, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.255803] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114550, 'name': CloneVM_Task, 'duration_secs': 1.846225} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.256230] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Created linked-clone VM from snapshot [ 1016.257191] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cff53ec-e79b-4611-9608-d3155ccd6f03 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.267453] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Uploading image e85cd8c4-7ac5-4359-bcda-17d18e3e60f9 {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1016.271171] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e248afc0-6637-487b-9510-8c22bafb82e1 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.464s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1016.301254] env[65522]: DEBUG oslo_vmware.rw_handles [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1016.301254] env[65522]: value = "vm-994888" [ 1016.301254] env[65522]: _type = "VirtualMachine" [ 1016.301254] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1016.301682] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f909caa9-e2e7-40ba-9682-fd9cff532971 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.312905] env[65522]: DEBUG oslo_vmware.rw_handles [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lease: (returnval){ [ 1016.312905] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523509d1-6d55-d960-19f0-fd7b59bb1c39" [ 1016.312905] env[65522]: _type = "HttpNfcLease" [ 1016.312905] env[65522]: } obtained for exporting VM: (result){ [ 1016.312905] env[65522]: value = "vm-994888" [ 1016.312905] env[65522]: _type = "VirtualMachine" [ 1016.312905] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1016.313345] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the lease: (returnval){ [ 1016.313345] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523509d1-6d55-d960-19f0-fd7b59bb1c39" [ 1016.313345] env[65522]: _type = "HttpNfcLease" [ 1016.313345] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1016.321057] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1016.321057] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523509d1-6d55-d960-19f0-fd7b59bb1c39" [ 1016.321057] env[65522]: _type = "HttpNfcLease" [ 1016.321057] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1016.375453] env[65522]: DEBUG nova.network.neutron [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Successfully created port: 1495d429-ea16-4882-990a-c5ccea69d0f8 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1016.420720] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527cc913-1c9e-3532-0db6-2cf6a05a4a2e, 'name': SearchDatastore_Task, 'duration_secs': 0.017801} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.421757] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ce31bc8-0342-4770-8f3a-61498b81f3ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.429955] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 1016.429955] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527127de-0950-56b2-5ba4-7f8d69ea1912" [ 1016.429955] env[65522]: _type = "Task" [ 1016.429955] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.444159] env[65522]: DEBUG oslo_concurrency.lockutils [None req-91ea8339-08a5-4cf7-9536-a16afec49825 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1016.445347] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527127de-0950-56b2-5ba4-7f8d69ea1912, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.445621] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 20.464s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1016.487822] env[65522]: WARNING neutronclient.v2_0.client [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1016.488600] env[65522]: WARNING openstack [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.489332] env[65522]: WARNING openstack [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.514914] env[65522]: DEBUG oslo_concurrency.lockutils [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Releasing lock "refresh_cache-5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1016.515608] env[65522]: DEBUG nova.compute.manager [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Received event network-changed-ed6e9224-69f8-46fd-a71f-27d8bc6eaec3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1016.516078] env[65522]: DEBUG nova.compute.manager [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Refreshing instance network info cache due to event network-changed-ed6e9224-69f8-46fd-a71f-27d8bc6eaec3. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1016.516706] env[65522]: DEBUG oslo_concurrency.lockutils [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Acquiring lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.517073] env[65522]: DEBUG oslo_concurrency.lockutils [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Acquired lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1016.517476] env[65522]: DEBUG nova.network.neutron [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Refreshing network info cache for port ed6e9224-69f8-46fd-a71f-27d8bc6eaec3 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1016.612720] env[65522]: DEBUG nova.network.neutron [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1016.648512] env[65522]: DEBUG oslo_vmware.api [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114552, 'name': PowerOffVM_Task, 'duration_secs': 0.519249} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.648857] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1016.649254] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1016.649367] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1200a0e4-6dba-4876-9bea-3c553f6456fd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.734375] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1016.734611] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1016.734799] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Deleting the datastore file [datastore2] ca964440-5375-4aff-8b45-96fbe829dd16 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1016.735113] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78d72c0b-090e-4709-89cc-a611f9f1e03e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.754669] env[65522]: DEBUG oslo_vmware.api [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for the task: (returnval){ [ 1016.754669] env[65522]: value = "task-5114555" [ 1016.754669] env[65522]: _type = "Task" [ 1016.754669] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.765127] env[65522]: DEBUG oslo_vmware.api [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114555, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.823240] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1016.823240] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523509d1-6d55-d960-19f0-fd7b59bb1c39" [ 1016.823240] env[65522]: _type = "HttpNfcLease" [ 1016.823240] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1016.823240] env[65522]: DEBUG oslo_vmware.rw_handles [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1016.823240] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523509d1-6d55-d960-19f0-fd7b59bb1c39" [ 1016.823240] env[65522]: _type = "HttpNfcLease" [ 1016.823240] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1016.824415] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e92b4bf6-dead-440f-87e1-ae78b62a90f5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.832384] env[65522]: DEBUG oslo_vmware.rw_handles [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c3946b-3fae-c66d-a2cb-8f70a88b38ae/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1016.832686] env[65522]: DEBUG oslo_vmware.rw_handles [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c3946b-3fae-c66d-a2cb-8f70a88b38ae/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1016.943548] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527127de-0950-56b2-5ba4-7f8d69ea1912, 'name': SearchDatastore_Task, 'duration_secs': 0.022148} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.943897] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1016.944246] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7/5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1016.944559] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-53ab9fe5-2108-4665-8c83-b3c63b56ed22 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.962762] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 1016.962762] env[65522]: value = "task-5114556" [ 1016.962762] env[65522]: _type = "Task" [ 1016.962762] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.970571] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-255f5cda-dea7-4dd4-b3ce-9001c97eab15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.981689] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114556, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.988713] env[65522]: DEBUG nova.compute.manager [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1017.021202] env[65522]: WARNING neutronclient.v2_0.client [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1017.021986] env[65522]: WARNING openstack [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1017.022366] env[65522]: WARNING openstack [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1017.193457] env[65522]: DEBUG nova.virt.hardware [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1017.193736] env[65522]: DEBUG nova.virt.hardware [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1017.193954] env[65522]: DEBUG nova.virt.hardware [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1017.194117] env[65522]: DEBUG nova.virt.hardware [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1017.194277] env[65522]: DEBUG nova.virt.hardware [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1017.194838] env[65522]: DEBUG nova.virt.hardware [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1017.194838] env[65522]: DEBUG nova.virt.hardware [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1017.194838] env[65522]: DEBUG nova.virt.hardware [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1017.195075] env[65522]: DEBUG nova.virt.hardware [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1017.195146] env[65522]: DEBUG nova.virt.hardware [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1017.195448] env[65522]: DEBUG nova.virt.hardware [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1017.196307] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-453e27b2-275a-4f65-b23d-b67c383f25da {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.209428] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5940e862-9857-46d4-a9f1-072a597635bf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.272204] env[65522]: DEBUG oslo_vmware.api [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Task: {'id': task-5114555, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.38166} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.272204] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1017.272427] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1017.272482] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1017.272957] env[65522]: INFO nova.compute.manager [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Took 1.66 seconds to destroy the instance on the hypervisor. [ 1017.272957] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1017.273440] env[65522]: DEBUG nova.compute.manager [-] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1017.273440] env[65522]: DEBUG nova.network.neutron [-] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1017.273562] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1017.274173] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1017.274522] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1017.332080] env[65522]: DEBUG nova.network.neutron [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1017.394931] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1017.477717] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114556, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.492068] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7c6f8218-602d-44f3-8012-de5a96972785 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1017.492253] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance ca964440-5375-4aff-8b45-96fbe829dd16 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1017.492374] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7617189c-a902-42e7-8165-0e7c4a1de06d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1017.492489] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 9c39b722-56ff-44fa-8f66-3e3432645a68 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1017.492831] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1017.492831] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 4aca4e87-b923-49e4-88d8-7ff51bcbe993 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1017.492831] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 79dd37b5-6b30-48ab-9f00-78214cbd132d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1017.492978] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 95a10a6f-8c6c-435d-a360-dff10eee0855 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1017.493175] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7701b92d-805e-4837-b900-326910b5eef5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1017.493175] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1017.493307] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 8859b051-8f75-4aad-b789-42662019d4c5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1017.493414] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 9f342b89-bde2-4c35-ae42-cfe1e6973b74 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1017.493519] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1017.493620] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance e7baf2a0-21dd-4610-8230-81b99b64856b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1017.493721] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1017.493824] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 43691901-7ac0-4523-a19c-eedf5ae4c040 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1017.815464] env[65522]: WARNING openstack [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1017.815926] env[65522]: WARNING openstack [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1017.834734] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dc6ea08-d458-4432-ac63-8d14efb2df07 req-79cb7f84-da6e-4ec1-a0ba-c16e4b30ca9e service nova] Releasing lock "refresh_cache-e7baf2a0-21dd-4610-8230-81b99b64856b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1017.835326] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquired lock "refresh_cache-e7baf2a0-21dd-4610-8230-81b99b64856b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1017.836026] env[65522]: DEBUG nova.network.neutron [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1017.916798] env[65522]: WARNING neutronclient.v2_0.client [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1017.917544] env[65522]: WARNING openstack [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1017.917900] env[65522]: WARNING openstack [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1017.957034] env[65522]: DEBUG nova.compute.manager [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Received event network-changed-2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1017.957034] env[65522]: DEBUG nova.compute.manager [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Refreshing instance network info cache due to event network-changed-2b2de45d-4c08-4bb4-810a-2f26266cbc20. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1017.957347] env[65522]: DEBUG oslo_concurrency.lockutils [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] Acquiring lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.957415] env[65522]: DEBUG oslo_concurrency.lockutils [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] Acquired lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1017.958045] env[65522]: DEBUG nova.network.neutron [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Refreshing network info cache for port 2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1017.990705] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114556, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.932085} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.991925] env[65522]: DEBUG nova.network.neutron [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Successfully updated port: 1495d429-ea16-4882-990a-c5ccea69d0f8 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1017.996922] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7/5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1017.996922] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1017.996922] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b61ffab7-a34e-4a27-bf3e-1605a38f9625 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.001079] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 99353848-2f0f-4388-9fcd-91e799342386 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1018.008057] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 1018.008057] env[65522]: value = "task-5114557" [ 1018.008057] env[65522]: _type = "Task" [ 1018.008057] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.024687] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114557, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.031665] env[65522]: DEBUG nova.network.neutron [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updated VIF entry in instance network info cache for port ed6e9224-69f8-46fd-a71f-27d8bc6eaec3. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1018.034229] env[65522]: DEBUG nova.network.neutron [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance_info_cache with network_info: [{"id": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "address": "fa:16:3e:c3:12:7d", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped6e9224-69", "ovs_interfaceid": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1018.195048] env[65522]: DEBUG nova.network.neutron [-] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1018.339536] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.340127] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.391948] env[65522]: DEBUG nova.network.neutron [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1018.413955] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.414316] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.461666] env[65522]: WARNING neutronclient.v2_0.client [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1018.462461] env[65522]: WARNING openstack [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.462822] env[65522]: WARNING openstack [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.491136] env[65522]: WARNING neutronclient.v2_0.client [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1018.491854] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.492515] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.501055] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "refresh_cache-43691901-7ac0-4523-a19c-eedf5ae4c040" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.501271] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "refresh_cache-43691901-7ac0-4523-a19c-eedf5ae4c040" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1018.501517] env[65522]: DEBUG nova.network.neutron [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1018.511908] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7a358133-88a5-4dd7-ab69-45b8172870a5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1018.525477] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114557, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096282} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.526081] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1018.527150] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55b928c-9e26-45a3-be66-90e5bc3a7280 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.546107] env[65522]: DEBUG oslo_concurrency.lockutils [req-e9d610b8-2d16-45d1-b4e3-4fc00d1f605f req-4f252030-6bcf-462e-bc32-f0a2589b4e71 service nova] Releasing lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1018.556021] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7/5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.566301] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2883df46-9988-49c3-8fec-e2be006b64b5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.591796] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 1018.591796] env[65522]: value = "task-5114558" [ 1018.591796] env[65522]: _type = "Task" [ 1018.591796] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.603849] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114558, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.660366] env[65522]: WARNING openstack [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.660921] env[65522]: WARNING openstack [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.678076] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.678657] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.697628] env[65522]: INFO nova.compute.manager [-] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Took 1.42 seconds to deallocate network for instance. [ 1018.731459] env[65522]: WARNING neutronclient.v2_0.client [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1018.732277] env[65522]: WARNING openstack [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.732812] env[65522]: WARNING openstack [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.755419] env[65522]: WARNING neutronclient.v2_0.client [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1018.756080] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.756428] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.777279] env[65522]: DEBUG oslo_vmware.rw_handles [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e6a733-8a95-e3df-a94e-c34a55047673/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1018.778647] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1c93a1-9357-454c-aec7-e6e02054c3b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.793705] env[65522]: DEBUG oslo_vmware.rw_handles [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e6a733-8a95-e3df-a94e-c34a55047673/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1018.794266] env[65522]: ERROR oslo_vmware.rw_handles [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e6a733-8a95-e3df-a94e-c34a55047673/disk-0.vmdk due to incomplete transfer. [ 1018.794657] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5eede5a2-4b3a-432e-8bcb-cf99d8902b12 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.807259] env[65522]: DEBUG oslo_vmware.rw_handles [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e6a733-8a95-e3df-a94e-c34a55047673/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1018.807563] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Uploaded image b0c7a88b-7872-412f-9204-61eaa49a2c35 to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1018.810920] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1018.811482] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c59743f0-cd31-4346-81cf-9a04bea23970 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.826332] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1018.826332] env[65522]: value = "task-5114559" [ 1018.826332] env[65522]: _type = "Task" [ 1018.826332] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.837160] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114559, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.848345] env[65522]: DEBUG nova.network.neutron [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updated VIF entry in instance network info cache for port 2b2de45d-4c08-4bb4-810a-2f26266cbc20. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1018.848834] env[65522]: DEBUG nova.network.neutron [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updating instance_info_cache with network_info: [{"id": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "address": "fa:16:3e:e9:bd:f6", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b2de45d-4c", "ovs_interfaceid": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1018.928807] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.929363] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.005477] env[65522]: WARNING openstack [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.005873] env[65522]: WARNING openstack [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.017203] env[65522]: WARNING neutronclient.v2_0.client [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1019.017468] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.017870] env[65522]: WARNING openstack [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.027019] env[65522]: DEBUG nova.compute.manager [req-2e3a4c2b-dfe6-4a96-b77e-37e80843021b req-cdfa61ea-4137-4efc-85f0-72f3b908d077 service nova] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Received event network-vif-plugged-1495d429-ea16-4882-990a-c5ccea69d0f8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1019.027116] env[65522]: DEBUG oslo_concurrency.lockutils [req-2e3a4c2b-dfe6-4a96-b77e-37e80843021b req-cdfa61ea-4137-4efc-85f0-72f3b908d077 service nova] Acquiring lock "43691901-7ac0-4523-a19c-eedf5ae4c040-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1019.027461] env[65522]: DEBUG oslo_concurrency.lockutils [req-2e3a4c2b-dfe6-4a96-b77e-37e80843021b req-cdfa61ea-4137-4efc-85f0-72f3b908d077 service nova] Lock "43691901-7ac0-4523-a19c-eedf5ae4c040-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1019.027560] env[65522]: DEBUG oslo_concurrency.lockutils [req-2e3a4c2b-dfe6-4a96-b77e-37e80843021b req-cdfa61ea-4137-4efc-85f0-72f3b908d077 service nova] Lock "43691901-7ac0-4523-a19c-eedf5ae4c040-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1019.027693] env[65522]: DEBUG nova.compute.manager [req-2e3a4c2b-dfe6-4a96-b77e-37e80843021b req-cdfa61ea-4137-4efc-85f0-72f3b908d077 service nova] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] No waiting events found dispatching network-vif-plugged-1495d429-ea16-4882-990a-c5ccea69d0f8 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1019.027858] env[65522]: WARNING nova.compute.manager [req-2e3a4c2b-dfe6-4a96-b77e-37e80843021b req-cdfa61ea-4137-4efc-85f0-72f3b908d077 service nova] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Received unexpected event network-vif-plugged-1495d429-ea16-4882-990a-c5ccea69d0f8 for instance with vm_state building and task_state spawning. [ 1019.028853] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance b412d3e5-0536-413c-9519-5f85c7647cdf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1019.029144] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1019.029277] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3200MB phys_disk=100GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '14', 'num_vm_active': '11', 'num_task_None': '8', 'num_os_type_None': '14', 'num_proj_fa11b46d9fe144f391233e6eb9c819d7': '2', 'io_workload': '3', 'num_task_deleting': '1', 'num_proj_9b8112ba49034e6fabcb3fbbd46edf41': '1', 'num_proj_91597233ae9c44c094f4c32d90332fa6': '1', 'num_proj_a80f979f3dc0477e9462b47f7aa87f14': '1', 'num_proj_9614f18e01594bf083e619c11760fe36': '2', 'num_task_shelving_image_uploading': '1', 'num_proj_1546bbdca07c443d84abea1155cfde37': '1', 'num_proj_065915a22e8b4ae09628cc71c7a188a0': '1', 'num_task_image_uploading': '1', 'num_proj_3ea0fab7de6c47eeaf3bdde413ae3901': '1', 'num_proj_0d1239b79ae94cceb89ae7a8bd57da08': '1', 'num_proj_068c2387de8c406194d9b1762c7292a7': '1', 'num_vm_building': '3', 'num_task_spawning': '2', 'num_proj_9db9e23ad1b745afa496de2deae83865': '1', 'num_proj_6ecccb656b0d4c96b40b200cdcddbad5': '1', 'num_task_block_device_mapping': '1'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1019.047488] env[65522]: DEBUG nova.network.neutron [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1019.087048] env[65522]: WARNING openstack [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.087048] env[65522]: WARNING openstack [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.108431] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.122033] env[65522]: DEBUG nova.network.neutron [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Updating instance_info_cache with network_info: [{"id": "44e3c08b-10b8-44fb-9e3e-1d2f77752195", "address": "fa:16:3e:60:e7:97", "network": {"id": "7def9dc4-2c9b-46b0-a2dc-56ad5343516c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-371441755", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.220", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44e3c08b-10", "ovs_interfaceid": "44e3c08b-10b8-44fb-9e3e-1d2f77752195", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "417e342f-fd78-422e-8ae2-05691d9ff64c", "address": "fa:16:3e:82:1c:b6", "network": {"id": "146ada64-89cb-4cde-81fd-a41d1a06a47c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-678002121", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.112", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap417e342f-fd", "ovs_interfaceid": "417e342f-fd78-422e-8ae2-05691d9ff64c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7052a26d-df5e-4222-be00-5bf53feb656f", "address": "fa:16:3e:f7:9f:e3", "network": {"id": "7def9dc4-2c9b-46b0-a2dc-56ad5343516c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-371441755", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.169", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7052a26d-df", "ovs_interfaceid": "7052a26d-df5e-4222-be00-5bf53feb656f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1019.165278] env[65522]: WARNING neutronclient.v2_0.client [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1019.166409] env[65522]: WARNING openstack [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.168747] env[65522]: WARNING openstack [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.206167] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1019.320638] env[65522]: DEBUG nova.network.neutron [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Updating instance_info_cache with network_info: [{"id": "1495d429-ea16-4882-990a-c5ccea69d0f8", "address": "fa:16:3e:22:2d:8c", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1495d429-ea", "ovs_interfaceid": "1495d429-ea16-4882-990a-c5ccea69d0f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1019.341136] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114559, 'name': Destroy_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.352388] env[65522]: DEBUG oslo_concurrency.lockutils [req-25740484-3724-4aa3-ae52-44ea9134a732 req-b3a66721-738a-44f6-acbe-df4759eb4dda service nova] Releasing lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1019.359518] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3384cd2-5aef-420c-b336-64e750254acb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.372266] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117d3c15-5f6b-4ac2-9efc-502203c75370 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.410305] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad04f63-41dc-465c-9347-45db4cb728a4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.421219] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7697f80-d6cf-4894-87e5-aaaf503dd854 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.439879] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.605724] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114558, 'name': ReconfigVM_Task, 'duration_secs': 0.782182} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.606032] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7/5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.606742] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e67abd4-840f-484e-a472-d5a30727beb0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.614948] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 1019.614948] env[65522]: value = "task-5114560" [ 1019.614948] env[65522]: _type = "Task" [ 1019.614948] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.624389] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114560, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.629135] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Releasing lock "refresh_cache-e7baf2a0-21dd-4610-8230-81b99b64856b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1019.629636] env[65522]: DEBUG nova.compute.manager [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Instance network_info: |[{"id": "44e3c08b-10b8-44fb-9e3e-1d2f77752195", "address": "fa:16:3e:60:e7:97", "network": {"id": "7def9dc4-2c9b-46b0-a2dc-56ad5343516c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-371441755", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.220", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44e3c08b-10", "ovs_interfaceid": "44e3c08b-10b8-44fb-9e3e-1d2f77752195", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "417e342f-fd78-422e-8ae2-05691d9ff64c", "address": "fa:16:3e:82:1c:b6", "network": {"id": "146ada64-89cb-4cde-81fd-a41d1a06a47c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-678002121", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.112", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap417e342f-fd", "ovs_interfaceid": "417e342f-fd78-422e-8ae2-05691d9ff64c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7052a26d-df5e-4222-be00-5bf53feb656f", "address": "fa:16:3e:f7:9f:e3", "network": {"id": "7def9dc4-2c9b-46b0-a2dc-56ad5343516c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-371441755", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.169", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7052a26d-df", "ovs_interfaceid": "7052a26d-df5e-4222-be00-5bf53feb656f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1019.630153] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:60:e7:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e2153f70-3d14-42ab-8bb3-be78296dd3b8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '44e3c08b-10b8-44fb-9e3e-1d2f77752195', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:1c:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60badc2d-69d2-467d-a92e-98511f5cb0b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '417e342f-fd78-422e-8ae2-05691d9ff64c', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:9f:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e2153f70-3d14-42ab-8bb3-be78296dd3b8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7052a26d-df5e-4222-be00-5bf53feb656f', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1019.643462] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Creating folder: Project (9db9e23ad1b745afa496de2deae83865). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1019.643819] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd7d2bf2-e271-46b9-83fb-654b44a89261 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.655785] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Created folder: Project (9db9e23ad1b745afa496de2deae83865) in parent group-v994660. [ 1019.655967] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Creating folder: Instances. Parent ref: group-v994890. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1019.656245] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c9335058-b91f-4f70-96bb-700f9f94e189 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.673834] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Created folder: Instances in parent group-v994890. [ 1019.674128] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1019.674350] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1019.674570] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f40e05ea-b3cb-49ec-86e6-7d3d049cda74 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.700084] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1019.700084] env[65522]: value = "task-5114563" [ 1019.700084] env[65522]: _type = "Task" [ 1019.700084] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.708816] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114563, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.775314] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1019.775548] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1019.828525] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "refresh_cache-43691901-7ac0-4523-a19c-eedf5ae4c040" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1019.828958] env[65522]: DEBUG nova.compute.manager [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Instance network_info: |[{"id": "1495d429-ea16-4882-990a-c5ccea69d0f8", "address": "fa:16:3e:22:2d:8c", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1495d429-ea", "ovs_interfaceid": "1495d429-ea16-4882-990a-c5ccea69d0f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1019.829555] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:2d:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10b81051-1eb1-406b-888c-4548c470c77e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1495d429-ea16-4882-990a-c5ccea69d0f8', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1019.837750] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1019.842311] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1019.843030] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-edd7c7f2-b815-43fd-9494-59b197ec336c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.866802] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114559, 'name': Destroy_Task, 'duration_secs': 0.843039} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.868467] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Destroyed the VM [ 1019.868586] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1019.868778] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1019.868778] env[65522]: value = "task-5114564" [ 1019.868778] env[65522]: _type = "Task" [ 1019.868778] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.868988] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-73ab0e16-7cd7-445e-a1ec-4ba0dcc3dfa0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.879877] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114564, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.881973] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1019.881973] env[65522]: value = "task-5114565" [ 1019.881973] env[65522]: _type = "Task" [ 1019.881973] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.891454] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114565, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.942860] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1019.990222] env[65522]: DEBUG nova.compute.manager [req-750bf1b2-41c1-4dbf-b973-8ddf68a2ce57 req-e41d633f-0895-4296-85b1-a188dc00d08e service nova] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Received event network-vif-deleted-ede8edfa-842b-482d-8e7b-34bde13a02f8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1020.126028] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114560, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.213080] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114563, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.278464] env[65522]: DEBUG nova.compute.manager [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1020.383537] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114564, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.393838] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114565, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.448924] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1020.449213] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.004s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1020.449491] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.344s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1020.451265] env[65522]: INFO nova.compute.claims [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1020.626687] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114560, 'name': Rename_Task, 'duration_secs': 0.91521} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.626992] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1020.627282] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-511d8cc4-0951-4371-8e0a-0775b32e5cc0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.637066] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 1020.637066] env[65522]: value = "task-5114566" [ 1020.637066] env[65522]: _type = "Task" [ 1020.637066] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.647585] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114566, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.713205] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114563, 'name': CreateVM_Task, 'duration_secs': 0.686818} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.713397] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1020.713950] env[65522]: WARNING neutronclient.v2_0.client [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.714319] env[65522]: WARNING neutronclient.v2_0.client [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.714581] env[65522]: WARNING neutronclient.v2_0.client [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.714898] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.715063] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1020.715396] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1020.715668] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b92fadaf-e7f5-4f91-953f-1bae74085395 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.721504] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1020.721504] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52697f42-f629-b1a9-0968-5e905685c308" [ 1020.721504] env[65522]: _type = "Task" [ 1020.721504] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.730782] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52697f42-f629-b1a9-0968-5e905685c308, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.809513] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1020.882553] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114564, 'name': CreateVM_Task, 'duration_secs': 0.892022} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.882938] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1020.883278] env[65522]: WARNING neutronclient.v2_0.client [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.883652] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.883804] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1020.884128] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1020.884443] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6c542f6-3d36-4098-9ab9-fb03cb79aedb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.894675] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114565, 'name': RemoveSnapshot_Task, 'duration_secs': 0.805793} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.895930] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1020.896234] env[65522]: DEBUG nova.compute.manager [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1020.896576] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1020.896576] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fb2eab-b144-b8c5-c5bc-ee5952002b91" [ 1020.896576] env[65522]: _type = "Task" [ 1020.896576] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.897340] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89aed31-85f9-4491-a0fe-50499d531906 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.911872] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fb2eab-b144-b8c5-c5bc-ee5952002b91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.049099] env[65522]: DEBUG nova.compute.manager [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Received event network-changed-1495d429-ea16-4882-990a-c5ccea69d0f8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1021.049315] env[65522]: DEBUG nova.compute.manager [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Refreshing instance network info cache due to event network-changed-1495d429-ea16-4882-990a-c5ccea69d0f8. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1021.049525] env[65522]: DEBUG oslo_concurrency.lockutils [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] Acquiring lock "refresh_cache-43691901-7ac0-4523-a19c-eedf5ae4c040" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.049662] env[65522]: DEBUG oslo_concurrency.lockutils [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] Acquired lock "refresh_cache-43691901-7ac0-4523-a19c-eedf5ae4c040" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1021.049819] env[65522]: DEBUG nova.network.neutron [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Refreshing network info cache for port 1495d429-ea16-4882-990a-c5ccea69d0f8 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1021.147492] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114566, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.233209] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52697f42-f629-b1a9-0968-5e905685c308, 'name': SearchDatastore_Task, 'duration_secs': 0.020507} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.233548] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1021.233836] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.234061] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.234175] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1021.234356] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.234649] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7730d393-54c7-4da2-98e7-f32e01ff5f6d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.246546] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.246729] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1021.247574] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82e6b0db-a426-4b17-ac52-a3424dd74ce4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.254925] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1021.254925] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5230d130-9971-8555-4251-455b98e107c0" [ 1021.254925] env[65522]: _type = "Task" [ 1021.254925] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.268033] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5230d130-9971-8555-4251-455b98e107c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.415088] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fb2eab-b144-b8c5-c5bc-ee5952002b91, 'name': SearchDatastore_Task, 'duration_secs': 0.030973} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.415992] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1021.416366] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.416684] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.416892] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1021.417178] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.417796] env[65522]: INFO nova.compute.manager [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Shelve offloading [ 1021.420818] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7af88138-aa84-4707-8523-aef3dd3c3673 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.433453] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.433760] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1021.434926] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-442b26e1-9885-41a5-890c-c003106da93e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.442613] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1021.442613] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a4a671-f4e7-f1d6-93b1-6f57e8d81bf2" [ 1021.442613] env[65522]: _type = "Task" [ 1021.442613] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.452555] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a4a671-f4e7-f1d6-93b1-6f57e8d81bf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.553025] env[65522]: WARNING neutronclient.v2_0.client [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1021.553989] env[65522]: WARNING openstack [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1021.554766] env[65522]: WARNING openstack [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1021.656970] env[65522]: DEBUG oslo_vmware.api [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114566, 'name': PowerOnVM_Task, 'duration_secs': 0.887243} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.661846] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1021.662092] env[65522]: INFO nova.compute.manager [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Took 9.61 seconds to spawn the instance on the hypervisor. [ 1021.662279] env[65522]: DEBUG nova.compute.manager [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1021.663368] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd2e1c9-6ea5-43da-9d9c-270e549a262c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.705243] env[65522]: WARNING openstack [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1021.705720] env[65522]: WARNING openstack [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1021.769527] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5230d130-9971-8555-4251-455b98e107c0, 'name': SearchDatastore_Task, 'duration_secs': 0.014764} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.770601] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1daf6f61-fd6f-4372-9aed-d2b8634c8eb5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.777632] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1021.777632] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521b1c76-690c-dbe0-c150-fa0a3dcc5a2e" [ 1021.777632] env[65522]: _type = "Task" [ 1021.777632] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.793608] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521b1c76-690c-dbe0-c150-fa0a3dcc5a2e, 'name': SearchDatastore_Task, 'duration_secs': 0.013033} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.793897] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1021.794183] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] e7baf2a0-21dd-4610-8230-81b99b64856b/e7baf2a0-21dd-4610-8230-81b99b64856b.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1021.794465] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f13c85a2-5955-45ad-af72-c1fd4d1c65ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.816865] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1021.816865] env[65522]: value = "task-5114567" [ 1021.816865] env[65522]: _type = "Task" [ 1021.816865] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.826454] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114567, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.826454] env[65522]: WARNING neutronclient.v2_0.client [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1021.826454] env[65522]: WARNING openstack [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1021.826988] env[65522]: WARNING openstack [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1021.840697] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e181080d-9add-4dd4-90de-df6110195940 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.849425] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f13e2f-8d8a-4ca3-87df-bb71f85c39e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.887074] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-062888a2-27f8-4dce-8b5c-f8215244bb0b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.896601] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80c878b-03ed-4b9b-8fad-f6d76bd14d43 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.912272] env[65522]: DEBUG nova.compute.provider_tree [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.924366] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.924664] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99435f16-6d9a-4e3c-b6b3-1d58e079ca55 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.931376] env[65522]: DEBUG nova.network.neutron [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Updated VIF entry in instance network info cache for port 1495d429-ea16-4882-990a-c5ccea69d0f8. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1021.931740] env[65522]: DEBUG nova.network.neutron [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Updating instance_info_cache with network_info: [{"id": "1495d429-ea16-4882-990a-c5ccea69d0f8", "address": "fa:16:3e:22:2d:8c", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1495d429-ea", "ovs_interfaceid": "1495d429-ea16-4882-990a-c5ccea69d0f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1021.934605] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1021.934605] env[65522]: value = "task-5114568" [ 1021.934605] env[65522]: _type = "Task" [ 1021.934605] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.949074] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1021.949281] env[65522]: DEBUG nova.compute.manager [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1021.950472] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ca416b-507b-40e9-acf6-ea3077bff6ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.960420] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a4a671-f4e7-f1d6-93b1-6f57e8d81bf2, 'name': SearchDatastore_Task, 'duration_secs': 0.019767} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.963054] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.963054] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1021.963054] env[65522]: DEBUG nova.network.neutron [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1021.964112] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59adfdcd-9401-4cdb-bb97-123c006b96cb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.970654] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1021.970654] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52aa119d-09ba-0dac-03da-c679e30cca9f" [ 1021.970654] env[65522]: _type = "Task" [ 1021.970654] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.980508] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52aa119d-09ba-0dac-03da-c679e30cca9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.185536] env[65522]: INFO nova.compute.manager [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Took 35.90 seconds to build instance. [ 1022.323074] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114567, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.415294] env[65522]: DEBUG nova.scheduler.client.report [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1022.436084] env[65522]: DEBUG oslo_concurrency.lockutils [req-b607d475-3d03-4d52-b712-e97bbc67677c req-b272ef0a-a0b4-4dde-b49e-aa747d8cd417 service nova] Releasing lock "refresh_cache-43691901-7ac0-4523-a19c-eedf5ae4c040" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1022.467536] env[65522]: WARNING neutronclient.v2_0.client [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1022.469059] env[65522]: WARNING openstack [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.469059] env[65522]: WARNING openstack [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.490065] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52aa119d-09ba-0dac-03da-c679e30cca9f, 'name': SearchDatastore_Task, 'duration_secs': 0.027644} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.490065] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1022.490065] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 43691901-7ac0-4523-a19c-eedf5ae4c040/43691901-7ac0-4523-a19c-eedf5ae4c040.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1022.490404] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-33d87362-b439-4ed6-aef1-80c3b604f4de {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.502756] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1022.502756] env[65522]: value = "task-5114569" [ 1022.502756] env[65522]: _type = "Task" [ 1022.502756] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.515415] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114569, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.622966] env[65522]: WARNING openstack [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.623492] env[65522]: WARNING openstack [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.688154] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8089e55c-026e-438e-beb3-cd8a14a6dfd7 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.431s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1022.694261] env[65522]: WARNING neutronclient.v2_0.client [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1022.695055] env[65522]: WARNING openstack [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1022.695558] env[65522]: WARNING openstack [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1022.798882] env[65522]: DEBUG nova.network.neutron [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Updating instance_info_cache with network_info: [{"id": "bd432f78-055d-4cd1-b502-8e058700c03a", "address": "fa:16:3e:38:8a:d1", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd432f78-05", "ovs_interfaceid": "bd432f78-055d-4cd1-b502-8e058700c03a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1022.822252] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114567, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.992332} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.822540] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] e7baf2a0-21dd-4610-8230-81b99b64856b/e7baf2a0-21dd-4610-8230-81b99b64856b.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1022.822763] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1022.823065] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc195aa7-0535-4d51-96e7-080f4909ac82 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.831346] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1022.831346] env[65522]: value = "task-5114570" [ 1022.831346] env[65522]: _type = "Task" [ 1022.831346] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.841220] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114570, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.920609] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.471s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1022.921370] env[65522]: DEBUG nova.compute.manager [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1022.925055] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.616s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1022.927283] env[65522]: INFO nova.compute.claims [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.013947] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114569, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.302535] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1023.302983] env[65522]: WARNING neutronclient.v2_0.client [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1023.303710] env[65522]: WARNING openstack [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1023.304084] env[65522]: WARNING openstack [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1023.310426] env[65522]: WARNING neutronclient.v2_0.client [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1023.345912] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114570, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089555} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.348036] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1023.348036] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0946718-8b6c-4426-b331-c2798fa47c60 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.380356] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] e7baf2a0-21dd-4610-8230-81b99b64856b/e7baf2a0-21dd-4610-8230-81b99b64856b.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.384270] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b7cc710-75d9-433e-a343-f06a4e21a018 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.410230] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1023.410230] env[65522]: value = "task-5114571" [ 1023.410230] env[65522]: _type = "Task" [ 1023.410230] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.427178] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114571, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.432308] env[65522]: DEBUG nova.compute.utils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1023.437841] env[65522]: DEBUG nova.compute.manager [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1023.438269] env[65522]: DEBUG nova.network.neutron [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1023.438740] env[65522]: WARNING neutronclient.v2_0.client [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1023.439205] env[65522]: WARNING neutronclient.v2_0.client [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1023.439902] env[65522]: WARNING openstack [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1023.440322] env[65522]: WARNING openstack [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1023.494448] env[65522]: DEBUG nova.policy [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f010fe7784043aaa3d4d44b921c438f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91597233ae9c44c094f4c32d90332fa6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1023.518026] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114569, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.703012] env[65522]: DEBUG oslo_concurrency.lockutils [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "9c39b722-56ff-44fa-8f66-3e3432645a68" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1023.703321] env[65522]: DEBUG oslo_concurrency.lockutils [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "9c39b722-56ff-44fa-8f66-3e3432645a68" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1023.703538] env[65522]: DEBUG oslo_concurrency.lockutils [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "9c39b722-56ff-44fa-8f66-3e3432645a68-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1023.703718] env[65522]: DEBUG oslo_concurrency.lockutils [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "9c39b722-56ff-44fa-8f66-3e3432645a68-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1023.703881] env[65522]: DEBUG oslo_concurrency.lockutils [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "9c39b722-56ff-44fa-8f66-3e3432645a68-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1023.706685] env[65522]: INFO nova.compute.manager [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Terminating instance [ 1023.791618] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1023.792566] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7c0ccd-9e54-4d2d-bd95-b95542c21c17 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.800738] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1023.801062] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-881fa016-4771-4087-a50c-9dac3c800d5c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.875273] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1023.875569] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1023.875709] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleting the datastore file [datastore1] 79dd37b5-6b30-48ab-9f00-78214cbd132d {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1023.875959] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28287f6d-1e9f-4ce7-bec6-a26030872d15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.883967] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1023.883967] env[65522]: value = "task-5114573" [ 1023.883967] env[65522]: _type = "Task" [ 1023.883967] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.893613] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114573, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.920527] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114571, 'name': ReconfigVM_Task, 'duration_secs': 0.372427} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.920823] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Reconfigured VM instance instance-00000055 to attach disk [datastore1] e7baf2a0-21dd-4610-8230-81b99b64856b/e7baf2a0-21dd-4610-8230-81b99b64856b.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1023.921656] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-830a8e25-7e79-4450-b4de-62959abb9e13 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.928263] env[65522]: DEBUG nova.network.neutron [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Successfully created port: 18edb16a-a08b-4e97-9b0a-a138798ed559 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1023.932322] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1023.932322] env[65522]: value = "task-5114574" [ 1023.932322] env[65522]: _type = "Task" [ 1023.932322] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.935888] env[65522]: DEBUG nova.compute.manager [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1023.948882] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114574, 'name': Rename_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.031088] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114569, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.10333} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.033052] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 43691901-7ac0-4523-a19c-eedf5ae4c040/43691901-7ac0-4523-a19c-eedf5ae4c040.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1024.033052] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1024.033052] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cefd8740-8692-424b-b555-4f5d0e603459 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.041223] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1024.041223] env[65522]: value = "task-5114575" [ 1024.041223] env[65522]: _type = "Task" [ 1024.041223] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.055556] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114575, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.211144] env[65522]: DEBUG nova.compute.manager [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1024.211387] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1024.212337] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0973ccf1-7276-45c0-a508-a238674042db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.222237] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1024.222572] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1589ae93-7726-4039-81ee-e5c137e4eb35 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.232221] env[65522]: DEBUG oslo_vmware.api [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1024.232221] env[65522]: value = "task-5114576" [ 1024.232221] env[65522]: _type = "Task" [ 1024.232221] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.243456] env[65522]: DEBUG oslo_vmware.api [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114576, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.317173] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f25f51-1e1d-4e1a-a89f-98d0184bff71 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.331597] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4605d25c-a26d-4356-b5a5-dc7971daaa42 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.375632] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e76254-922e-423f-949a-ef8a61f021f2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.384536] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35da716-6e35-4dbc-9a36-122fd7b25dcf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.396593] env[65522]: DEBUG nova.compute.manager [req-a64364b3-26e9-4a69-a180-51598baaf98b req-7867179e-b351-4ef3-8af2-501ae469d298 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Received event network-vif-unplugged-bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1024.396842] env[65522]: DEBUG oslo_concurrency.lockutils [req-a64364b3-26e9-4a69-a180-51598baaf98b req-7867179e-b351-4ef3-8af2-501ae469d298 service nova] Acquiring lock "79dd37b5-6b30-48ab-9f00-78214cbd132d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1024.398195] env[65522]: DEBUG oslo_concurrency.lockutils [req-a64364b3-26e9-4a69-a180-51598baaf98b req-7867179e-b351-4ef3-8af2-501ae469d298 service nova] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1024.398793] env[65522]: DEBUG oslo_concurrency.lockutils [req-a64364b3-26e9-4a69-a180-51598baaf98b req-7867179e-b351-4ef3-8af2-501ae469d298 service nova] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1024.398793] env[65522]: DEBUG nova.compute.manager [req-a64364b3-26e9-4a69-a180-51598baaf98b req-7867179e-b351-4ef3-8af2-501ae469d298 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] No waiting events found dispatching network-vif-unplugged-bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1024.398793] env[65522]: WARNING nova.compute.manager [req-a64364b3-26e9-4a69-a180-51598baaf98b req-7867179e-b351-4ef3-8af2-501ae469d298 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Received unexpected event network-vif-unplugged-bd432f78-055d-4cd1-b502-8e058700c03a for instance with vm_state shelved and task_state shelving_offloading. [ 1024.412030] env[65522]: DEBUG nova.compute.provider_tree [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.413498] env[65522]: DEBUG oslo_vmware.api [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114573, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284304} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.414227] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1024.414445] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1024.414796] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1024.437618] env[65522]: INFO nova.scheduler.client.report [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleted allocations for instance 79dd37b5-6b30-48ab-9f00-78214cbd132d [ 1024.449793] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114574, 'name': Rename_Task, 'duration_secs': 0.193243} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.450108] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1024.450387] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-91890051-064a-4698-bfdb-c875aadb4735 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.458844] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1024.458844] env[65522]: value = "task-5114577" [ 1024.458844] env[65522]: _type = "Task" [ 1024.458844] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.468530] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114577, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.551711] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114575, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086359} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.552020] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1024.552942] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab4ba8e-2833-42a3-8bb9-1e9ec13c1928 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.579353] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 43691901-7ac0-4523-a19c-eedf5ae4c040/43691901-7ac0-4523-a19c-eedf5ae4c040.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1024.579776] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa73b539-ddbe-4c0c-b8e3-9d2d8d733bb5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.607882] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1024.607882] env[65522]: value = "task-5114578" [ 1024.607882] env[65522]: _type = "Task" [ 1024.607882] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.618710] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114578, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.743026] env[65522]: DEBUG oslo_vmware.api [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114576, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.915592] env[65522]: DEBUG nova.scheduler.client.report [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1024.945548] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1024.946744] env[65522]: DEBUG nova.compute.manager [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1024.978362] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114577, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.981513] env[65522]: DEBUG nova.virt.hardware [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1024.981513] env[65522]: DEBUG nova.virt.hardware [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1024.981513] env[65522]: DEBUG nova.virt.hardware [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1024.982267] env[65522]: DEBUG nova.virt.hardware [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1024.982267] env[65522]: DEBUG nova.virt.hardware [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1024.982267] env[65522]: DEBUG nova.virt.hardware [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1024.982267] env[65522]: DEBUG nova.virt.hardware [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1024.982395] env[65522]: DEBUG nova.virt.hardware [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1024.982787] env[65522]: DEBUG nova.virt.hardware [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1024.982787] env[65522]: DEBUG nova.virt.hardware [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1024.982787] env[65522]: DEBUG nova.virt.hardware [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1024.983714] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d45386-6773-4b18-92af-1129bc1e84b6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.996615] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f46ff6f-36c0-4d5d-bc73-e37b7af9b558 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.122339] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114578, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.244466] env[65522]: DEBUG oslo_vmware.api [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114576, 'name': PowerOffVM_Task, 'duration_secs': 0.551612} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.245173] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.245173] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1025.245520] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a4404d3-483c-466e-9df1-e706230db145 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.329280] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1025.329583] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1025.329826] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleting the datastore file [datastore2] 9c39b722-56ff-44fa-8f66-3e3432645a68 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.330923] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-950b48cd-e466-4de4-9430-e624194f73de {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.339516] env[65522]: DEBUG oslo_vmware.api [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1025.339516] env[65522]: value = "task-5114580" [ 1025.339516] env[65522]: _type = "Task" [ 1025.339516] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.350857] env[65522]: DEBUG oslo_vmware.api [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114580, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.421893] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.422027] env[65522]: DEBUG nova.compute.manager [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1025.427870] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.094s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1025.428142] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.430580] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.888s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1025.432073] env[65522]: INFO nova.compute.claims [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1025.470862] env[65522]: INFO nova.scheduler.client.report [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Deleted allocations for instance 7701b92d-805e-4837-b900-326910b5eef5 [ 1025.476076] env[65522]: DEBUG oslo_vmware.api [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114577, 'name': PowerOnVM_Task, 'duration_secs': 0.703041} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.480680] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1025.480680] env[65522]: INFO nova.compute.manager [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Took 15.91 seconds to spawn the instance on the hypervisor. [ 1025.480680] env[65522]: DEBUG nova.compute.manager [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1025.481531] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3398bdea-c18b-4841-9fdd-057a5dfd5139 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.591482] env[65522]: DEBUG nova.compute.manager [req-382ca97a-604b-44f2-b1a8-7d4cb1aa93e3 req-f1c3e7c5-ded9-4e56-b221-a7c26d738a76 service nova] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Received event network-vif-plugged-18edb16a-a08b-4e97-9b0a-a138798ed559 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1025.591482] env[65522]: DEBUG oslo_concurrency.lockutils [req-382ca97a-604b-44f2-b1a8-7d4cb1aa93e3 req-f1c3e7c5-ded9-4e56-b221-a7c26d738a76 service nova] Acquiring lock "99353848-2f0f-4388-9fcd-91e799342386-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1025.591712] env[65522]: DEBUG oslo_concurrency.lockutils [req-382ca97a-604b-44f2-b1a8-7d4cb1aa93e3 req-f1c3e7c5-ded9-4e56-b221-a7c26d738a76 service nova] Lock "99353848-2f0f-4388-9fcd-91e799342386-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1025.591942] env[65522]: DEBUG oslo_concurrency.lockutils [req-382ca97a-604b-44f2-b1a8-7d4cb1aa93e3 req-f1c3e7c5-ded9-4e56-b221-a7c26d738a76 service nova] Lock "99353848-2f0f-4388-9fcd-91e799342386-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.592790] env[65522]: DEBUG nova.compute.manager [req-382ca97a-604b-44f2-b1a8-7d4cb1aa93e3 req-f1c3e7c5-ded9-4e56-b221-a7c26d738a76 service nova] [instance: 99353848-2f0f-4388-9fcd-91e799342386] No waiting events found dispatching network-vif-plugged-18edb16a-a08b-4e97-9b0a-a138798ed559 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1025.592790] env[65522]: WARNING nova.compute.manager [req-382ca97a-604b-44f2-b1a8-7d4cb1aa93e3 req-f1c3e7c5-ded9-4e56-b221-a7c26d738a76 service nova] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Received unexpected event network-vif-plugged-18edb16a-a08b-4e97-9b0a-a138798ed559 for instance with vm_state building and task_state spawning. [ 1025.620019] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114578, 'name': ReconfigVM_Task, 'duration_secs': 0.777057} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.620189] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 43691901-7ac0-4523-a19c-eedf5ae4c040/43691901-7ac0-4523-a19c-eedf5ae4c040.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.620789] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ff4a91ad-ea02-478c-b671-a14ed3113d23 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.629403] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1025.629403] env[65522]: value = "task-5114581" [ 1025.629403] env[65522]: _type = "Task" [ 1025.629403] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.640763] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114581, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.649345] env[65522]: DEBUG nova.network.neutron [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Successfully updated port: 18edb16a-a08b-4e97-9b0a-a138798ed559 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1025.779994] env[65522]: DEBUG oslo_concurrency.lockutils [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1025.780381] env[65522]: DEBUG oslo_concurrency.lockutils [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1025.780606] env[65522]: DEBUG oslo_concurrency.lockutils [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1025.781819] env[65522]: DEBUG oslo_concurrency.lockutils [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1025.781819] env[65522]: DEBUG oslo_concurrency.lockutils [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.786378] env[65522]: INFO nova.compute.manager [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Terminating instance [ 1025.850634] env[65522]: DEBUG oslo_vmware.api [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114580, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.302211} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.850935] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1025.851159] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1025.851497] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1025.851605] env[65522]: INFO nova.compute.manager [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Took 1.64 seconds to destroy the instance on the hypervisor. [ 1025.851795] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1025.852376] env[65522]: DEBUG nova.compute.manager [-] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1025.852517] env[65522]: DEBUG nova.network.neutron [-] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1025.852709] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1025.853447] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1025.853554] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1025.900223] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1025.939167] env[65522]: DEBUG nova.compute.utils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1025.944797] env[65522]: DEBUG nova.compute.manager [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1025.945154] env[65522]: DEBUG nova.network.neutron [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1025.945535] env[65522]: WARNING neutronclient.v2_0.client [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1025.946423] env[65522]: WARNING neutronclient.v2_0.client [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1025.947183] env[65522]: WARNING openstack [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1025.947597] env[65522]: WARNING openstack [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1025.984631] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9965004e-a330-4b2d-8dbf-f7d567dd856b tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "7701b92d-805e-4837-b900-326910b5eef5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.957s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1026.005300] env[65522]: INFO nova.compute.manager [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Took 40.64 seconds to build instance. [ 1026.015693] env[65522]: DEBUG nova.policy [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8f8fa586d8c74bb9805a1148fa7ed037', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f93c790ff61543bd8e134bcf9cb20bb2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1026.140703] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114581, 'name': Rename_Task, 'duration_secs': 0.218761} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.141179] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1026.141367] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e64a8bc7-ccae-4bc3-9144-d7bcf0f8ab36 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.150119] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1026.150119] env[65522]: value = "task-5114582" [ 1026.150119] env[65522]: _type = "Task" [ 1026.150119] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.157284] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "refresh_cache-99353848-2f0f-4388-9fcd-91e799342386" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.157402] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "refresh_cache-99353848-2f0f-4388-9fcd-91e799342386" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1026.157584] env[65522]: DEBUG nova.network.neutron [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1026.168599] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114582, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.292456] env[65522]: DEBUG nova.compute.manager [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1026.293391] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1026.294623] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d239052a-f540-49d0-8d5a-7243fbe79bfd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.307782] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1026.308193] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f769a899-2ae8-4213-b42f-0f086d250f7b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.319909] env[65522]: DEBUG oslo_vmware.api [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 1026.319909] env[65522]: value = "task-5114583" [ 1026.319909] env[65522]: _type = "Task" [ 1026.319909] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.328987] env[65522]: DEBUG oslo_vmware.api [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114583, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.461871] env[65522]: DEBUG nova.compute.manager [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1026.468278] env[65522]: DEBUG nova.network.neutron [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Successfully created port: 0831bbe0-c665-4be1-932f-e4ed9c52eb4e {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1026.492291] env[65522]: DEBUG nova.compute.manager [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Received event network-changed-bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1026.492291] env[65522]: DEBUG nova.compute.manager [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Refreshing instance network info cache due to event network-changed-bd432f78-055d-4cd1-b502-8e058700c03a. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1026.492291] env[65522]: DEBUG oslo_concurrency.lockutils [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] Acquiring lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.492291] env[65522]: DEBUG oslo_concurrency.lockutils [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] Acquired lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1026.492291] env[65522]: DEBUG nova.network.neutron [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Refreshing network info cache for port bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1026.514890] env[65522]: DEBUG oslo_concurrency.lockutils [None req-18b894cd-bbf8-4b0a-b439-5f9ff8cc0297 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "e7baf2a0-21dd-4610-8230-81b99b64856b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.160s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1026.668041] env[65522]: WARNING openstack [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1026.668978] env[65522]: WARNING openstack [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1026.677563] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114582, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.681129] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "95a10a6f-8c6c-435d-a360-dff10eee0855" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1026.681458] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "95a10a6f-8c6c-435d-a360-dff10eee0855" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1026.681701] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "95a10a6f-8c6c-435d-a360-dff10eee0855-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1026.681921] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "95a10a6f-8c6c-435d-a360-dff10eee0855-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1026.682139] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "95a10a6f-8c6c-435d-a360-dff10eee0855-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1026.685656] env[65522]: INFO nova.compute.manager [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Terminating instance [ 1026.746498] env[65522]: DEBUG nova.network.neutron [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1026.829236] env[65522]: DEBUG oslo_vmware.api [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114583, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.844508] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1026.872730] env[65522]: WARNING openstack [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1026.873165] env[65522]: WARNING openstack [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1026.883267] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc32af6-111b-4bfe-a689-3421c79ec50c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.891899] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c099a2-9898-4484-b163-599f4d0e7e12 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.933896] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db12824-cba5-451a-aac8-9d869e11535e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.936788] env[65522]: DEBUG nova.network.neutron [-] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1026.949483] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c27d6c-96f3-443c-b2ba-f29a02dfc071 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.969430] env[65522]: DEBUG nova.compute.provider_tree [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.976123] env[65522]: WARNING neutronclient.v2_0.client [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1026.976841] env[65522]: WARNING openstack [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1026.977209] env[65522]: WARNING openstack [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.000731] env[65522]: WARNING neutronclient.v2_0.client [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1027.001623] env[65522]: WARNING openstack [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1027.001993] env[65522]: WARNING openstack [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.143708] env[65522]: DEBUG nova.network.neutron [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Updating instance_info_cache with network_info: [{"id": "18edb16a-a08b-4e97-9b0a-a138798ed559", "address": "fa:16:3e:f6:b9:dd", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18edb16a-a0", "ovs_interfaceid": "18edb16a-a08b-4e97-9b0a-a138798ed559", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1027.161893] env[65522]: DEBUG oslo_vmware.api [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114582, 'name': PowerOnVM_Task, 'duration_secs': 0.659417} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.162171] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1027.162321] env[65522]: INFO nova.compute.manager [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Took 10.17 seconds to spawn the instance on the hypervisor. [ 1027.162500] env[65522]: DEBUG nova.compute.manager [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1027.163320] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2403da2-ee0b-4307-b156-937da94cdc58 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.195542] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "refresh_cache-95a10a6f-8c6c-435d-a360-dff10eee0855" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.195542] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquired lock "refresh_cache-95a10a6f-8c6c-435d-a360-dff10eee0855" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1027.195542] env[65522]: DEBUG nova.network.neutron [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1027.321486] env[65522]: WARNING openstack [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1027.321929] env[65522]: WARNING openstack [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.339424] env[65522]: DEBUG oslo_vmware.api [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114583, 'name': PowerOffVM_Task, 'duration_secs': 0.571566} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.339737] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1027.339905] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1027.340202] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d94d112-1eba-4a7a-a5d6-ba800eeb54e2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.439966] env[65522]: INFO nova.compute.manager [-] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Took 1.59 seconds to deallocate network for instance. [ 1027.450019] env[65522]: DEBUG oslo_concurrency.lockutils [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "e7baf2a0-21dd-4610-8230-81b99b64856b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.450285] env[65522]: DEBUG oslo_concurrency.lockutils [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "e7baf2a0-21dd-4610-8230-81b99b64856b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.450495] env[65522]: DEBUG oslo_concurrency.lockutils [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "e7baf2a0-21dd-4610-8230-81b99b64856b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.450673] env[65522]: DEBUG oslo_concurrency.lockutils [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "e7baf2a0-21dd-4610-8230-81b99b64856b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.450836] env[65522]: DEBUG oslo_concurrency.lockutils [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "e7baf2a0-21dd-4610-8230-81b99b64856b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1027.453213] env[65522]: INFO nova.compute.manager [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Terminating instance [ 1027.471732] env[65522]: DEBUG nova.scheduler.client.report [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1027.488574] env[65522]: DEBUG nova.compute.manager [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1027.490988] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1027.491305] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1027.491807] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleting the datastore file [datastore1] 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1027.492837] env[65522]: WARNING neutronclient.v2_0.client [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1027.493608] env[65522]: WARNING openstack [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1027.494049] env[65522]: WARNING openstack [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.503460] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5110f4ff-1d1c-4e16-859a-45e3a31924f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.513221] env[65522]: DEBUG oslo_vmware.api [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for the task: (returnval){ [ 1027.513221] env[65522]: value = "task-5114585" [ 1027.513221] env[65522]: _type = "Task" [ 1027.513221] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.525205] env[65522]: DEBUG oslo_vmware.api [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114585, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.534983] env[65522]: DEBUG nova.virt.hardware [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1027.535521] env[65522]: DEBUG nova.virt.hardware [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1027.535718] env[65522]: DEBUG nova.virt.hardware [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1027.535918] env[65522]: DEBUG nova.virt.hardware [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1027.536124] env[65522]: DEBUG nova.virt.hardware [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1027.536301] env[65522]: DEBUG nova.virt.hardware [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1027.536573] env[65522]: DEBUG nova.virt.hardware [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1027.536767] env[65522]: DEBUG nova.virt.hardware [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1027.536980] env[65522]: DEBUG nova.virt.hardware [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1027.537174] env[65522]: DEBUG nova.virt.hardware [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1027.537403] env[65522]: DEBUG nova.virt.hardware [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1027.538421] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f780682-4ff7-4290-982a-4ec8d0cdd00d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.548997] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414e1c65-4227-4b9c-9d9d-66d7dfe9b42c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.626438] env[65522]: DEBUG nova.compute.manager [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Received event network-changed-18edb16a-a08b-4e97-9b0a-a138798ed559 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1027.626636] env[65522]: DEBUG nova.compute.manager [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Refreshing instance network info cache due to event network-changed-18edb16a-a08b-4e97-9b0a-a138798ed559. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1027.626821] env[65522]: DEBUG oslo_concurrency.lockutils [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] Acquiring lock "refresh_cache-99353848-2f0f-4388-9fcd-91e799342386" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.637477] env[65522]: DEBUG nova.network.neutron [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Updated VIF entry in instance network info cache for port bd432f78-055d-4cd1-b502-8e058700c03a. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1027.637827] env[65522]: DEBUG nova.network.neutron [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Updating instance_info_cache with network_info: [{"id": "bd432f78-055d-4cd1-b502-8e058700c03a", "address": "fa:16:3e:38:8a:d1", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapbd432f78-05", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1027.646543] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "refresh_cache-99353848-2f0f-4388-9fcd-91e799342386" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1027.647085] env[65522]: DEBUG nova.compute.manager [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Instance network_info: |[{"id": "18edb16a-a08b-4e97-9b0a-a138798ed559", "address": "fa:16:3e:f6:b9:dd", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18edb16a-a0", "ovs_interfaceid": "18edb16a-a08b-4e97-9b0a-a138798ed559", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1027.647245] env[65522]: DEBUG oslo_concurrency.lockutils [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] Acquired lock "refresh_cache-99353848-2f0f-4388-9fcd-91e799342386" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1027.647423] env[65522]: DEBUG nova.network.neutron [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Refreshing network info cache for port 18edb16a-a08b-4e97-9b0a-a138798ed559 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1027.648649] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:b9:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56136ef6-99d7-4562-9a9f-d66fec951c5c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '18edb16a-a08b-4e97-9b0a-a138798ed559', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1027.656799] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1027.658232] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1027.658901] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b492a21c-8011-4db5-b40c-0441e31a1d56 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.685701] env[65522]: INFO nova.compute.manager [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Took 38.73 seconds to build instance. [ 1027.688435] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1027.688435] env[65522]: value = "task-5114586" [ 1027.688435] env[65522]: _type = "Task" [ 1027.688435] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.698301] env[65522]: WARNING neutronclient.v2_0.client [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1027.698925] env[65522]: WARNING openstack [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1027.699347] env[65522]: WARNING openstack [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1027.706713] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114586, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.948415] env[65522]: DEBUG oslo_concurrency.lockutils [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.959426] env[65522]: DEBUG nova.compute.manager [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1027.959867] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1027.961149] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14dececb-ece5-43b4-b035-66230748674d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.974985] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1027.976079] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a9ad115-5632-41fb-8602-e7fc3ddbea81 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.983462] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.553s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1027.984334] env[65522]: DEBUG nova.compute.manager [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1027.992459] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.861s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.992663] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.002s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1027.996329] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.790s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.996699] env[65522]: DEBUG nova.objects.instance [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lazy-loading 'resources' on Instance uuid ca964440-5375-4aff-8b45-96fbe829dd16 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.002341] env[65522]: DEBUG nova.network.neutron [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1028.011186] env[65522]: DEBUG oslo_vmware.api [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1028.011186] env[65522]: value = "task-5114587" [ 1028.011186] env[65522]: _type = "Task" [ 1028.011186] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.026860] env[65522]: DEBUG oslo_vmware.api [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114587, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.032507] env[65522]: DEBUG oslo_vmware.api [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Task: {'id': task-5114585, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.351752} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.034015] env[65522]: INFO nova.scheduler.client.report [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Deleted allocations for instance 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83 [ 1028.035982] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1028.036813] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1028.036813] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1028.036813] env[65522]: INFO nova.compute.manager [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Took 1.74 seconds to destroy the instance on the hypervisor. [ 1028.037136] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1028.038794] env[65522]: DEBUG nova.compute.manager [-] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1028.038794] env[65522]: DEBUG nova.network.neutron [-] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1028.038794] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1028.039413] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.040037] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1028.087139] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1028.131734] env[65522]: DEBUG nova.network.neutron [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1028.140393] env[65522]: DEBUG oslo_concurrency.lockutils [req-8c39f354-47c3-4eda-862d-c0bffa206529 req-e53d6c8e-013d-42f8-acb0-9f43bdd300d2 service nova] Releasing lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1028.158051] env[65522]: WARNING neutronclient.v2_0.client [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1028.158838] env[65522]: WARNING openstack [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.159506] env[65522]: WARNING openstack [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1028.186452] env[65522]: DEBUG oslo_concurrency.lockutils [None req-61126e5c-c077-466e-b1cd-2fbad861fcab tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "43691901-7ac0-4523-a19c-eedf5ae4c040" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.246s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1028.199947] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114586, 'name': CreateVM_Task, 'duration_secs': 0.455356} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.200116] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1028.200665] env[65522]: WARNING neutronclient.v2_0.client [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1028.201100] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.201267] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1028.201624] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1028.202039] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02c7f39d-50b1-4a13-aaac-fb60073ee02d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.210519] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1028.210519] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526a2a2d-b853-689b-42b5-057904ae7d2d" [ 1028.210519] env[65522]: _type = "Task" [ 1028.210519] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.225811] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526a2a2d-b853-689b-42b5-057904ae7d2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.238141] env[65522]: DEBUG nova.network.neutron [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Successfully updated port: 0831bbe0-c665-4be1-932f-e4ed9c52eb4e {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1028.378294] env[65522]: DEBUG oslo_vmware.rw_handles [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c3946b-3fae-c66d-a2cb-8f70a88b38ae/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1028.379296] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc36acb-e439-413e-bee4-2ace9fbd4913 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.387680] env[65522]: DEBUG oslo_vmware.rw_handles [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c3946b-3fae-c66d-a2cb-8f70a88b38ae/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1028.388054] env[65522]: ERROR oslo_vmware.rw_handles [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c3946b-3fae-c66d-a2cb-8f70a88b38ae/disk-0.vmdk due to incomplete transfer. [ 1028.388100] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-69b5dd00-bd79-442f-b912-a2c36c4c49ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.396745] env[65522]: DEBUG oslo_vmware.rw_handles [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c3946b-3fae-c66d-a2cb-8f70a88b38ae/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1028.396961] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Uploaded image e85cd8c4-7ac5-4359-bcda-17d18e3e60f9 to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1028.401501] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1028.402520] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-43e4e7f4-194f-431d-9f2e-dc216f317b73 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.417067] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1028.417067] env[65522]: value = "task-5114588" [ 1028.417067] env[65522]: _type = "Task" [ 1028.417067] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.428100] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114588, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.431373] env[65522]: WARNING openstack [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.431373] env[65522]: WARNING openstack [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1028.492241] env[65522]: DEBUG nova.compute.utils [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1028.493833] env[65522]: DEBUG nova.compute.manager [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Not allocating networking since 'none' was specified. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 1028.516072] env[65522]: WARNING neutronclient.v2_0.client [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1028.516835] env[65522]: WARNING openstack [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.517209] env[65522]: WARNING openstack [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1028.530647] env[65522]: DEBUG oslo_vmware.api [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114587, 'name': PowerOffVM_Task, 'duration_secs': 0.268464} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.530890] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1028.531111] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1028.531390] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-be57fbe4-0644-4f85-a1b1-cbf3a775c93c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.546203] env[65522]: DEBUG nova.compute.manager [req-c1440ce0-e1fc-4427-8262-71954bbd3da2 req-f1385c60-7761-46ef-a598-e60058ee4913 service nova] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Received event network-vif-deleted-54206d03-4403-4bca-8ba8-1f4e88682cb6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1028.546755] env[65522]: DEBUG nova.compute.manager [req-c1440ce0-e1fc-4427-8262-71954bbd3da2 req-f1385c60-7761-46ef-a598-e60058ee4913 service nova] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Received event network-vif-deleted-15e24bea-7053-40cb-8a80-6782dcca755f {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1028.546850] env[65522]: INFO nova.compute.manager [req-c1440ce0-e1fc-4427-8262-71954bbd3da2 req-f1385c60-7761-46ef-a598-e60058ee4913 service nova] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Neutron deleted interface 15e24bea-7053-40cb-8a80-6782dcca755f; detaching it from the instance and deleting it from the info cache [ 1028.547229] env[65522]: DEBUG nova.network.neutron [req-c1440ce0-e1fc-4427-8262-71954bbd3da2 req-f1385c60-7761-46ef-a598-e60058ee4913 service nova] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1028.558995] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e00ca54-2dcc-469b-99a7-1f526aeb353b tempest-ServerAddressesNegativeTestJSON-709887233 tempest-ServerAddressesNegativeTestJSON-709887233-project-member] Lock "00fcbbcb-ef87-4318-8c6e-ce62feb9fd83" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.376s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1028.629430] env[65522]: DEBUG nova.network.neutron [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Updated VIF entry in instance network info cache for port 18edb16a-a08b-4e97-9b0a-a138798ed559. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1028.629824] env[65522]: DEBUG nova.network.neutron [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Updating instance_info_cache with network_info: [{"id": "18edb16a-a08b-4e97-9b0a-a138798ed559", "address": "fa:16:3e:f6:b9:dd", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18edb16a-a0", "ovs_interfaceid": "18edb16a-a08b-4e97-9b0a-a138798ed559", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1028.635388] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Releasing lock "refresh_cache-95a10a6f-8c6c-435d-a360-dff10eee0855" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1028.635772] env[65522]: DEBUG nova.compute.manager [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1028.635977] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1028.637045] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a2bb1b-074e-46b9-b0b5-ac6891a71d8b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.648586] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1028.649591] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-006a97f7-7eec-4ccb-9a69-ae41b50b2086 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.660950] env[65522]: DEBUG oslo_vmware.api [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 1028.660950] env[65522]: value = "task-5114590" [ 1028.660950] env[65522]: _type = "Task" [ 1028.660950] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.672400] env[65522]: DEBUG oslo_vmware.api [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114590, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.677260] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1028.677448] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1028.677659] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Deleting the datastore file [datastore1] e7baf2a0-21dd-4610-8230-81b99b64856b {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1028.677994] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-916e12c3-49fe-4c06-9983-3d348f04e687 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.686595] env[65522]: DEBUG oslo_vmware.api [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1028.686595] env[65522]: value = "task-5114591" [ 1028.686595] env[65522]: _type = "Task" [ 1028.686595] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.700154] env[65522]: DEBUG oslo_vmware.api [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114591, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.721940] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526a2a2d-b853-689b-42b5-057904ae7d2d, 'name': SearchDatastore_Task, 'duration_secs': 0.022186} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.722693] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1028.722938] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1028.723211] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.723353] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1028.723529] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1028.723817] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e722ce2-6131-4e1f-a33c-e3cd7927e591 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.735650] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1028.735856] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1028.736710] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17173b97-b68f-49dc-8e53-ddb101412aed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.745400] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "refresh_cache-7a358133-88a5-4dd7-ab69-45b8172870a5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.745648] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "refresh_cache-7a358133-88a5-4dd7-ab69-45b8172870a5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1028.746729] env[65522]: DEBUG nova.network.neutron [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1028.751022] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1028.751022] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52311e0f-c6e9-57b8-ba18-8de467548d0a" [ 1028.751022] env[65522]: _type = "Task" [ 1028.751022] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.759455] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52311e0f-c6e9-57b8-ba18-8de467548d0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.803631] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13da19f0-987c-48f3-839a-34674fb862eb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.813246] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f28287-e7e5-4de4-b039-ff498f0b8e34 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.847670] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b29533-8a69-4e48-8eb4-64bcecd5984b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.856038] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ecccc0-567f-426b-89e7-5b6a08a58f54 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.875267] env[65522]: DEBUG nova.compute.provider_tree [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.914377] env[65522]: DEBUG nova.network.neutron [-] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1028.928987] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114588, 'name': Destroy_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.996111] env[65522]: DEBUG nova.compute.manager [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1029.053476] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d5882c7f-08cc-4487-a852-5147e6c7188d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.064116] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d689aa-6b4f-4f2c-8709-d5b6fc2d0e72 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.110405] env[65522]: DEBUG nova.compute.manager [req-c1440ce0-e1fc-4427-8262-71954bbd3da2 req-f1385c60-7761-46ef-a598-e60058ee4913 service nova] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Detach interface failed, port_id=15e24bea-7053-40cb-8a80-6782dcca755f, reason: Instance 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1029.134984] env[65522]: DEBUG oslo_concurrency.lockutils [req-b8e23d47-366c-4867-af9c-2f01ae9d7326 req-68a0f8fc-fa3b-4bc9-9726-5053f28aae09 service nova] Releasing lock "refresh_cache-99353848-2f0f-4388-9fcd-91e799342386" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1029.173103] env[65522]: DEBUG oslo_vmware.api [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114590, 'name': PowerOffVM_Task, 'duration_secs': 0.164175} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.173382] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1029.173583] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1029.173918] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-182d47f0-5752-4796-8614-36aedec00f6d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.200387] env[65522]: DEBUG oslo_vmware.api [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114591, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.369883} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.201800] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.202149] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1029.202371] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1029.202553] env[65522]: INFO nova.compute.manager [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Took 1.24 seconds to destroy the instance on the hypervisor. [ 1029.202815] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1029.203045] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1029.203224] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1029.203415] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Deleting the datastore file [datastore1] 95a10a6f-8c6c-435d-a360-dff10eee0855 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1029.203658] env[65522]: DEBUG nova.compute.manager [-] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1029.203756] env[65522]: DEBUG nova.network.neutron [-] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1029.204332] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.205147] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.205403] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.213239] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-41c3e58d-fd2f-4fbe-8736-03e2a5ff5e6e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.222038] env[65522]: DEBUG oslo_vmware.api [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for the task: (returnval){ [ 1029.222038] env[65522]: value = "task-5114593" [ 1029.222038] env[65522]: _type = "Task" [ 1029.222038] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.232779] env[65522]: DEBUG oslo_vmware.api [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114593, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.252873] env[65522]: WARNING openstack [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.253412] env[65522]: WARNING openstack [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.268988] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52311e0f-c6e9-57b8-ba18-8de467548d0a, 'name': SearchDatastore_Task, 'duration_secs': 0.01865} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.269870] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43857822-cc81-4972-b313-ab602a8f726a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.276411] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1029.276411] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c0fb1b-e4a6-585f-cd45-d4cf36525bdd" [ 1029.276411] env[65522]: _type = "Task" [ 1029.276411] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.277612] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.291574] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c0fb1b-e4a6-585f-cd45-d4cf36525bdd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.307813] env[65522]: DEBUG nova.network.neutron [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1029.380143] env[65522]: DEBUG nova.scheduler.client.report [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1029.388016] env[65522]: WARNING openstack [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.388457] env[65522]: WARNING openstack [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.423147] env[65522]: INFO nova.compute.manager [-] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Took 1.38 seconds to deallocate network for instance. [ 1029.431653] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114588, 'name': Destroy_Task, 'duration_secs': 0.712564} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.432177] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Destroyed the VM [ 1029.433393] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1029.433393] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-53afe079-e0d1-42b7-837d-bbb4a5a3e165 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.441284] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1029.441284] env[65522]: value = "task-5114594" [ 1029.441284] env[65522]: _type = "Task" [ 1029.441284] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.450772] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114594, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.458439] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "43691901-7ac0-4523-a19c-eedf5ae4c040" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1029.458726] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "43691901-7ac0-4523-a19c-eedf5ae4c040" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1029.458990] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "43691901-7ac0-4523-a19c-eedf5ae4c040-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1029.459208] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "43691901-7ac0-4523-a19c-eedf5ae4c040-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1029.459761] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "43691901-7ac0-4523-a19c-eedf5ae4c040-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.466810] env[65522]: INFO nova.compute.manager [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Terminating instance [ 1029.733336] env[65522]: DEBUG oslo_vmware.api [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Task: {'id': task-5114593, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150497} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.733689] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.733882] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1029.734075] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1029.734502] env[65522]: INFO nova.compute.manager [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1029.734502] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1029.734689] env[65522]: DEBUG nova.compute.manager [-] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1029.734792] env[65522]: DEBUG nova.network.neutron [-] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1029.735057] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.735609] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.735865] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.797853] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c0fb1b-e4a6-585f-cd45-d4cf36525bdd, 'name': SearchDatastore_Task, 'duration_secs': 0.014173} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.798396] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1029.798840] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 99353848-2f0f-4388-9fcd-91e799342386/99353848-2f0f-4388-9fcd-91e799342386.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1029.799303] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12928c01-4e3a-44d9-8402-2afb26126373 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.808783] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1029.808783] env[65522]: value = "task-5114595" [ 1029.808783] env[65522]: _type = "Task" [ 1029.808783] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.821330] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114595, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.861369] env[65522]: WARNING neutronclient.v2_0.client [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.862053] env[65522]: WARNING openstack [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.862437] env[65522]: WARNING openstack [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.887067] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.890s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.889063] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.080s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1029.890672] env[65522]: INFO nova.compute.claims [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1029.909109] env[65522]: DEBUG nova.network.neutron [-] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1029.909109] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.929679] env[65522]: INFO nova.scheduler.client.report [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Deleted allocations for instance ca964440-5375-4aff-8b45-96fbe829dd16 [ 1029.937932] env[65522]: DEBUG oslo_concurrency.lockutils [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1029.951343] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114594, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.975787] env[65522]: DEBUG nova.compute.manager [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1029.976046] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1029.979119] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dffcc173-0a38-45dc-a44e-8ad9d3614cee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.989146] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1029.990869] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01aadf61-5c18-43b7-b97c-878e7ec9528d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.999278] env[65522]: DEBUG oslo_vmware.api [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1029.999278] env[65522]: value = "task-5114596" [ 1029.999278] env[65522]: _type = "Task" [ 1029.999278] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.010356] env[65522]: DEBUG nova.compute.manager [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1030.012654] env[65522]: DEBUG oslo_vmware.api [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114596, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.049237] env[65522]: DEBUG nova.virt.hardware [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1030.049522] env[65522]: DEBUG nova.virt.hardware [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1030.049695] env[65522]: DEBUG nova.virt.hardware [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1030.049960] env[65522]: DEBUG nova.virt.hardware [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1030.050045] env[65522]: DEBUG nova.virt.hardware [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1030.050193] env[65522]: DEBUG nova.virt.hardware [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1030.050413] env[65522]: DEBUG nova.virt.hardware [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.050594] env[65522]: DEBUG nova.virt.hardware [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1030.050763] env[65522]: DEBUG nova.virt.hardware [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1030.050965] env[65522]: DEBUG nova.virt.hardware [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1030.051160] env[65522]: DEBUG nova.virt.hardware [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1030.052527] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f032321e-778f-4a1a-bb3a-134ee3d313b6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.065869] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4dd4097-ac0c-4775-9be8-769c9e255a6b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.089634] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1030.096852] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Creating folder: Project (b30d3cdbb77d4438ab379c2e2597c6d2). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1030.097208] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d87b4a9-f524-4b55-850f-945fe60913f4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.112607] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Created folder: Project (b30d3cdbb77d4438ab379c2e2597c6d2) in parent group-v994660. [ 1030.112780] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Creating folder: Instances. Parent ref: group-v994895. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1030.113031] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-25aa7cd4-d30a-4f47-849a-8c96cb400c08 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.130846] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Created folder: Instances in parent group-v994895. [ 1030.131383] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1030.131876] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1030.132165] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e9f9f7e-6f01-46fd-8836-8a2564ff1e21 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.156829] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1030.156829] env[65522]: value = "task-5114599" [ 1030.156829] env[65522]: _type = "Task" [ 1030.156829] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.172312] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114599, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.212394] env[65522]: DEBUG nova.compute.manager [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Received event network-vif-plugged-0831bbe0-c665-4be1-932f-e4ed9c52eb4e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1030.212394] env[65522]: DEBUG oslo_concurrency.lockutils [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] Acquiring lock "7a358133-88a5-4dd7-ab69-45b8172870a5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1030.213318] env[65522]: DEBUG oslo_concurrency.lockutils [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] Lock "7a358133-88a5-4dd7-ab69-45b8172870a5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1030.213318] env[65522]: DEBUG oslo_concurrency.lockutils [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] Lock "7a358133-88a5-4dd7-ab69-45b8172870a5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1030.213318] env[65522]: DEBUG nova.compute.manager [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] No waiting events found dispatching network-vif-plugged-0831bbe0-c665-4be1-932f-e4ed9c52eb4e {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1030.213318] env[65522]: WARNING nova.compute.manager [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Received unexpected event network-vif-plugged-0831bbe0-c665-4be1-932f-e4ed9c52eb4e for instance with vm_state building and task_state spawning. [ 1030.214076] env[65522]: DEBUG nova.compute.manager [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Received event network-changed-0831bbe0-c665-4be1-932f-e4ed9c52eb4e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1030.214076] env[65522]: DEBUG nova.compute.manager [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Refreshing instance network info cache due to event network-changed-0831bbe0-c665-4be1-932f-e4ed9c52eb4e. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1030.214076] env[65522]: DEBUG oslo_concurrency.lockutils [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] Acquiring lock "refresh_cache-7a358133-88a5-4dd7-ab69-45b8172870a5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.277192] env[65522]: DEBUG nova.network.neutron [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Updating instance_info_cache with network_info: [{"id": "0831bbe0-c665-4be1-932f-e4ed9c52eb4e", "address": "fa:16:3e:51:5f:69", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0831bbe0-c6", "ovs_interfaceid": "0831bbe0-c665-4be1-932f-e4ed9c52eb4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1030.325235] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114595, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.412467] env[65522]: DEBUG nova.network.neutron [-] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1030.446538] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec447d12-c9b3-4283-9a01-5d756e4b0224 tempest-ServerRescueNegativeTestJSON-611467277 tempest-ServerRescueNegativeTestJSON-611467277-project-member] Lock "ca964440-5375-4aff-8b45-96fbe829dd16" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.338s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1030.457118] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114594, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.509159] env[65522]: DEBUG oslo_vmware.api [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114596, 'name': PowerOffVM_Task, 'duration_secs': 0.246693} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.509634] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1030.509731] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1030.510407] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2bdae09a-c037-452f-8e8f-76ffd9943878 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.583386] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1030.583386] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1030.583386] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleting the datastore file [datastore2] 43691901-7ac0-4523-a19c-eedf5ae4c040 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1030.583549] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f739794-8f60-49e3-8a1e-8cb63427ee24 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.593058] env[65522]: DEBUG oslo_vmware.api [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1030.593058] env[65522]: value = "task-5114601" [ 1030.593058] env[65522]: _type = "Task" [ 1030.593058] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.603271] env[65522]: DEBUG oslo_vmware.api [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114601, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.668704] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114599, 'name': CreateVM_Task, 'duration_secs': 0.480441} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.668704] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1030.669274] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.669991] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1030.669991] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1030.671367] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2957ebbd-807e-41b7-9d94-96a2d973d751 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.676668] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1030.676668] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b3add4-54c9-052a-f89d-ddac21cfc034" [ 1030.676668] env[65522]: _type = "Task" [ 1030.676668] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.686358] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b3add4-54c9-052a-f89d-ddac21cfc034, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.694080] env[65522]: DEBUG nova.network.neutron [-] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1030.737829] env[65522]: DEBUG nova.compute.manager [req-535782a0-03ec-4538-ba98-de9bcc85b336 req-c8f03c47-5637-4b97-a123-005ac384d027 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Received event network-vif-deleted-417e342f-fd78-422e-8ae2-05691d9ff64c {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1030.738104] env[65522]: INFO nova.compute.manager [req-535782a0-03ec-4538-ba98-de9bcc85b336 req-c8f03c47-5637-4b97-a123-005ac384d027 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Neutron deleted interface 417e342f-fd78-422e-8ae2-05691d9ff64c; detaching it from the instance and deleting it from the info cache [ 1030.738412] env[65522]: DEBUG nova.network.neutron [req-535782a0-03ec-4538-ba98-de9bcc85b336 req-c8f03c47-5637-4b97-a123-005ac384d027 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Updating instance_info_cache with network_info: [{"id": "44e3c08b-10b8-44fb-9e3e-1d2f77752195", "address": "fa:16:3e:60:e7:97", "network": {"id": "7def9dc4-2c9b-46b0-a2dc-56ad5343516c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-371441755", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.220", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44e3c08b-10", "ovs_interfaceid": "44e3c08b-10b8-44fb-9e3e-1d2f77752195", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7052a26d-df5e-4222-be00-5bf53feb656f", "address": "fa:16:3e:f7:9f:e3", "network": {"id": "7def9dc4-2c9b-46b0-a2dc-56ad5343516c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-371441755", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.169", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7052a26d-df", "ovs_interfaceid": "7052a26d-df5e-4222-be00-5bf53feb656f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1030.783070] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "refresh_cache-7a358133-88a5-4dd7-ab69-45b8172870a5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1030.783070] env[65522]: DEBUG nova.compute.manager [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Instance network_info: |[{"id": "0831bbe0-c665-4be1-932f-e4ed9c52eb4e", "address": "fa:16:3e:51:5f:69", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0831bbe0-c6", "ovs_interfaceid": "0831bbe0-c665-4be1-932f-e4ed9c52eb4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1030.783070] env[65522]: DEBUG oslo_concurrency.lockutils [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] Acquired lock "refresh_cache-7a358133-88a5-4dd7-ab69-45b8172870a5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1030.783070] env[65522]: DEBUG nova.network.neutron [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Refreshing network info cache for port 0831bbe0-c665-4be1-932f-e4ed9c52eb4e {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1030.784448] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:5f:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0831bbe0-c665-4be1-932f-e4ed9c52eb4e', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1030.798319] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1030.798428] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1030.798648] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fc72a708-bc27-43bc-9431-237280a9cbaa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.825686] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114595, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.567993} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.827105] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 99353848-2f0f-4388-9fcd-91e799342386/99353848-2f0f-4388-9fcd-91e799342386.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1030.827342] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1030.827871] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1030.827871] env[65522]: value = "task-5114602" [ 1030.827871] env[65522]: _type = "Task" [ 1030.827871] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.828083] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-806a5f5f-696f-41f8-8e7a-99a9cc51e660 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.839313] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114602, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.840805] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1030.840805] env[65522]: value = "task-5114603" [ 1030.840805] env[65522]: _type = "Task" [ 1030.840805] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.914626] env[65522]: INFO nova.compute.manager [-] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Took 1.18 seconds to deallocate network for instance. [ 1030.961204] env[65522]: DEBUG oslo_vmware.api [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114594, 'name': RemoveSnapshot_Task, 'duration_secs': 1.02524} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.961204] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1030.961204] env[65522]: INFO nova.compute.manager [None req-7906ef77-5626-497a-8701-0b287b065f51 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Took 19.84 seconds to snapshot the instance on the hypervisor. [ 1031.108391] env[65522]: DEBUG oslo_vmware.api [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114601, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173424} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.108682] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1031.109189] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1031.109189] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1031.109330] env[65522]: INFO nova.compute.manager [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1031.109516] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1031.109711] env[65522]: DEBUG nova.compute.manager [-] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1031.109804] env[65522]: DEBUG nova.network.neutron [-] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1031.110061] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1031.110748] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.111529] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.157015] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1031.189096] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b3add4-54c9-052a-f89d-ddac21cfc034, 'name': SearchDatastore_Task, 'duration_secs': 0.011906} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.192998] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.193188] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1031.193451] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.193762] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1031.193853] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1031.194433] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9da5c3c0-1465-4c7f-ae71-0f3a8359c605 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.198446] env[65522]: INFO nova.compute.manager [-] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Took 1.99 seconds to deallocate network for instance. [ 1031.210793] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1031.211142] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1031.212122] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-465aa98d-6aec-4f2f-83ec-7674284781d6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.220881] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1031.220881] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e7b7cb-4e52-9b7a-8680-a57c14eabe92" [ 1031.220881] env[65522]: _type = "Task" [ 1031.220881] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.226102] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9611aa05-410e-4be9-b12d-81ae16b0d138 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.240054] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e7b7cb-4e52-9b7a-8680-a57c14eabe92, 'name': SearchDatastore_Task} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.242302] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74893385-b41f-4100-868d-3480023aebd6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.244877] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-63de7a19-9c7b-4eab-86d2-8a2c53b09612 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.250052] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad1dddd-458a-439c-a914-e2bda3fbbeba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.288304] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1031.288304] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52437dc3-a4ff-687e-c1bb-d523f677ca0f" [ 1031.288304] env[65522]: _type = "Task" [ 1031.288304] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.292368] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95592840-1ce3-4bb7-ad9f-b2c0677c234f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.302658] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f637ac-e9e2-4e54-9fb6-4aabf7cae313 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.316395] env[65522]: WARNING neutronclient.v2_0.client [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1031.317087] env[65522]: WARNING openstack [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.317463] env[65522]: WARNING openstack [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.331649] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca5f64a8-b5e9-4b86-83c3-2e981b5f5687 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.343991] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52437dc3-a4ff-687e-c1bb-d523f677ca0f, 'name': SearchDatastore_Task, 'duration_secs': 0.01257} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.356351] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.356647] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] b412d3e5-0536-413c-9519-5f85c7647cdf/b412d3e5-0536-413c-9519-5f85c7647cdf.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1031.357052] env[65522]: DEBUG nova.compute.manager [req-535782a0-03ec-4538-ba98-de9bcc85b336 req-c8f03c47-5637-4b97-a123-005ac384d027 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Detach interface failed, port_id=417e342f-fd78-422e-8ae2-05691d9ff64c, reason: Instance e7baf2a0-21dd-4610-8230-81b99b64856b could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1031.357268] env[65522]: DEBUG nova.compute.manager [req-535782a0-03ec-4538-ba98-de9bcc85b336 req-c8f03c47-5637-4b97-a123-005ac384d027 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Received event network-vif-deleted-7052a26d-df5e-4222-be00-5bf53feb656f {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1031.357425] env[65522]: INFO nova.compute.manager [req-535782a0-03ec-4538-ba98-de9bcc85b336 req-c8f03c47-5637-4b97-a123-005ac384d027 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Neutron deleted interface 7052a26d-df5e-4222-be00-5bf53feb656f; detaching it from the instance and deleting it from the info cache [ 1031.357691] env[65522]: DEBUG nova.network.neutron [req-535782a0-03ec-4538-ba98-de9bcc85b336 req-c8f03c47-5637-4b97-a123-005ac384d027 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Updating instance_info_cache with network_info: [{"id": "44e3c08b-10b8-44fb-9e3e-1d2f77752195", "address": "fa:16:3e:60:e7:97", "network": {"id": "7def9dc4-2c9b-46b0-a2dc-56ad5343516c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-371441755", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.220", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44e3c08b-10", "ovs_interfaceid": "44e3c08b-10b8-44fb-9e3e-1d2f77752195", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1031.371663] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8006250a-fd10-42d4-ba68-5753ae4b7614 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.374791] env[65522]: DEBUG nova.compute.provider_tree [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1031.380104] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114602, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.386967] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114603, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.183622} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.388785] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1031.389274] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1031.389274] env[65522]: value = "task-5114604" [ 1031.389274] env[65522]: _type = "Task" [ 1031.389274] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.390522] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1adb04-8e08-4ae3-aacf-551c77fa39bb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.422037] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 99353848-2f0f-4388-9fcd-91e799342386/99353848-2f0f-4388-9fcd-91e799342386.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.426726] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7de735fd-dc71-40ed-9f2e-c8f46c65be3e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.443165] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114604, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.446998] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1031.455770] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1031.455770] env[65522]: value = "task-5114605" [ 1031.455770] env[65522]: _type = "Task" [ 1031.455770] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.468670] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "f1ce2d19-347a-4830-8501-8cceb87d4bf8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1031.468670] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "f1ce2d19-347a-4830-8501-8cceb87d4bf8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1031.471837] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114605, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.713116] env[65522]: DEBUG oslo_concurrency.lockutils [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1031.739217] env[65522]: WARNING openstack [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.739664] env[65522]: WARNING openstack [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.842670] env[65522]: WARNING neutronclient.v2_0.client [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1031.843609] env[65522]: WARNING openstack [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1031.844167] env[65522]: WARNING openstack [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1031.867032] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114602, 'name': CreateVM_Task, 'duration_secs': 0.564329} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.867032] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-25bd035d-27ff-4d3b-8741-e1cefe47cbf8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.868804] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1031.869841] env[65522]: WARNING neutronclient.v2_0.client [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1031.870429] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.870693] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1031.872021] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1031.872021] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96e1726a-c16b-49cb-aa54-6386aae86e9d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.878458] env[65522]: DEBUG nova.scheduler.client.report [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1031.885218] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1031.885218] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bcae6d-df03-10dc-dcbd-5087a21ec955" [ 1031.885218] env[65522]: _type = "Task" [ 1031.885218] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.889244] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bccfa7f3-108b-447d-8eba-60da0c29fad2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.928132] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114604, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.932059] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bcae6d-df03-10dc-dcbd-5087a21ec955, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.946296] env[65522]: DEBUG nova.compute.manager [req-535782a0-03ec-4538-ba98-de9bcc85b336 req-c8f03c47-5637-4b97-a123-005ac384d027 service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Detach interface failed, port_id=7052a26d-df5e-4222-be00-5bf53feb656f, reason: Instance e7baf2a0-21dd-4610-8230-81b99b64856b could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1031.947814] env[65522]: DEBUG nova.network.neutron [-] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1031.969065] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114605, 'name': ReconfigVM_Task, 'duration_secs': 0.387186} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.970263] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 99353848-2f0f-4388-9fcd-91e799342386/99353848-2f0f-4388-9fcd-91e799342386.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1031.971224] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5a5bf3cd-2d76-4abe-8479-781de5073b9f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.973827] env[65522]: DEBUG nova.compute.manager [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1031.984130] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1031.984130] env[65522]: value = "task-5114606" [ 1031.984130] env[65522]: _type = "Task" [ 1031.984130] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.995733] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114606, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.005530] env[65522]: DEBUG nova.network.neutron [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Updated VIF entry in instance network info cache for port 0831bbe0-c665-4be1-932f-e4ed9c52eb4e. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1032.005907] env[65522]: DEBUG nova.network.neutron [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Updating instance_info_cache with network_info: [{"id": "0831bbe0-c665-4be1-932f-e4ed9c52eb4e", "address": "fa:16:3e:51:5f:69", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0831bbe0-c6", "ovs_interfaceid": "0831bbe0-c665-4be1-932f-e4ed9c52eb4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1032.259086] env[65522]: DEBUG nova.compute.manager [req-89cb9eab-a64e-405c-9311-e1a97cee9cb4 req-1d94fd98-f78f-4251-81f4-3f5b4aa7dd6c service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Received event network-vif-deleted-44e3c08b-10b8-44fb-9e3e-1d2f77752195 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1032.259339] env[65522]: INFO nova.compute.manager [req-89cb9eab-a64e-405c-9311-e1a97cee9cb4 req-1d94fd98-f78f-4251-81f4-3f5b4aa7dd6c service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Neutron deleted interface 44e3c08b-10b8-44fb-9e3e-1d2f77752195; detaching it from the instance and deleting it from the info cache [ 1032.259485] env[65522]: DEBUG nova.network.neutron [req-89cb9eab-a64e-405c-9311-e1a97cee9cb4 req-1d94fd98-f78f-4251-81f4-3f5b4aa7dd6c service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1032.387477] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1032.388034] env[65522]: DEBUG nova.compute.manager [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1032.391810] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.446s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1032.391810] env[65522]: DEBUG nova.objects.instance [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lazy-loading 'resources' on Instance uuid 79dd37b5-6b30-48ab-9f00-78214cbd132d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.420127] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bcae6d-df03-10dc-dcbd-5087a21ec955, 'name': SearchDatastore_Task, 'duration_secs': 0.06067} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.423280] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1032.423761] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1032.424096] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.424303] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1032.424908] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1032.424908] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114604, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.603825} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.425389] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ee4bba8-0ded-4d89-8c3b-f901559fb98f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.427564] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] b412d3e5-0536-413c-9519-5f85c7647cdf/b412d3e5-0536-413c-9519-5f85c7647cdf.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1032.427564] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1032.427833] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c0606c3-3ebe-425c-84ac-9a9d9466aa4e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.439363] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1032.439363] env[65522]: value = "task-5114607" [ 1032.439363] env[65522]: _type = "Task" [ 1032.439363] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.443271] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1032.443457] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1032.444745] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b2f1b79-e1e6-4d2f-b755-f3b6a69cfe04 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.452784] env[65522]: INFO nova.compute.manager [-] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Took 1.34 seconds to deallocate network for instance. [ 1032.452784] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114607, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.457345] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1032.457345] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525d8161-4394-0414-9878-e774fda90687" [ 1032.457345] env[65522]: _type = "Task" [ 1032.457345] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.472324] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525d8161-4394-0414-9878-e774fda90687, 'name': SearchDatastore_Task, 'duration_secs': 0.011794} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.473563] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-632121fe-2525-4983-a65d-39a284aeb8ed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.482540] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1032.482540] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dab996-ad2a-848f-bf39-9f2d2524761a" [ 1032.482540] env[65522]: _type = "Task" [ 1032.482540] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.499074] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114606, 'name': Rename_Task, 'duration_secs': 0.177966} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.502621] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1032.502955] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dab996-ad2a-848f-bf39-9f2d2524761a, 'name': SearchDatastore_Task, 'duration_secs': 0.010607} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.504187] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1032.504751] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e5e842c7-bfa3-4220-9c1f-97840be34faa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.506620] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1032.506753] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 7a358133-88a5-4dd7-ab69-45b8172870a5/7a358133-88a5-4dd7-ab69-45b8172870a5.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1032.507031] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-510aabcf-32a4-4a1c-aa69-f041604a7a1f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.509889] env[65522]: DEBUG oslo_concurrency.lockutils [req-baf2d85f-1f67-4bfa-b5a7-117353eb6da5 req-41b0c91e-c1d3-4c7b-a366-307965208f28 service nova] Releasing lock "refresh_cache-7a358133-88a5-4dd7-ab69-45b8172870a5" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1032.516050] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1032.516050] env[65522]: value = "task-5114608" [ 1032.516050] env[65522]: _type = "Task" [ 1032.516050] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.516283] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1032.516283] env[65522]: value = "task-5114609" [ 1032.516283] env[65522]: _type = "Task" [ 1032.516283] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.528923] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114608, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.532532] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114609, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.763427] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cefde7f8-1917-4ebd-b959-8433da7d712b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.780153] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485658cf-d88c-4803-9632-e8124449e895 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.826804] env[65522]: DEBUG nova.compute.manager [req-89cb9eab-a64e-405c-9311-e1a97cee9cb4 req-1d94fd98-f78f-4251-81f4-3f5b4aa7dd6c service nova] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Detach interface failed, port_id=44e3c08b-10b8-44fb-9e3e-1d2f77752195, reason: Instance e7baf2a0-21dd-4610-8230-81b99b64856b could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1032.827071] env[65522]: DEBUG nova.compute.manager [req-89cb9eab-a64e-405c-9311-e1a97cee9cb4 req-1d94fd98-f78f-4251-81f4-3f5b4aa7dd6c service nova] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Received event network-vif-deleted-1495d429-ea16-4882-990a-c5ccea69d0f8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1032.897068] env[65522]: DEBUG nova.compute.utils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1032.897068] env[65522]: DEBUG nova.compute.manager [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1032.897068] env[65522]: DEBUG nova.network.neutron [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1032.897677] env[65522]: WARNING neutronclient.v2_0.client [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1032.898206] env[65522]: WARNING neutronclient.v2_0.client [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1032.899036] env[65522]: WARNING openstack [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1032.899558] env[65522]: WARNING openstack [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1032.909359] env[65522]: DEBUG nova.objects.instance [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lazy-loading 'numa_topology' on Instance uuid 79dd37b5-6b30-48ab-9f00-78214cbd132d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.911843] env[65522]: DEBUG nova.compute.manager [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1032.957203] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114607, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075623} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.957498] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1032.958595] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08026eef-ad96-43d4-a9b0-a53b7f6877bb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.966238] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1032.970839] env[65522]: DEBUG nova.policy [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b491939798e4481fb433ffb81f366b25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '068c2387de8c406194d9b1762c7292a7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1032.991296] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] b412d3e5-0536-413c-9519-5f85c7647cdf/b412d3e5-0536-413c-9519-5f85c7647cdf.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.992573] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46ff3a42-10d6-47da-8159-8844cb496c0a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.027080] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1033.027080] env[65522]: value = "task-5114610" [ 1033.027080] env[65522]: _type = "Task" [ 1033.027080] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.044242] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114608, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.044423] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114609, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.048409] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114610, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.388255] env[65522]: DEBUG nova.network.neutron [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Successfully created port: d443e903-a2fa-4c64-b185-81fe54af293a {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1033.416298] env[65522]: DEBUG nova.objects.base [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Object Instance<79dd37b5-6b30-48ab-9f00-78214cbd132d> lazy-loaded attributes: resources,numa_topology {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1033.535542] env[65522]: DEBUG oslo_vmware.api [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114608, 'name': PowerOnVM_Task, 'duration_secs': 0.578486} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.535785] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114609, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.580719} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.539040] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1033.539040] env[65522]: INFO nova.compute.manager [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Took 8.59 seconds to spawn the instance on the hypervisor. [ 1033.539188] env[65522]: DEBUG nova.compute.manager [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1033.539468] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 7a358133-88a5-4dd7-ab69-45b8172870a5/7a358133-88a5-4dd7-ab69-45b8172870a5.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1033.539662] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1033.542687] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94578f1e-84d7-4a16-b2c6-7cb063876d3e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.545307] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f153cfd4-4c82-4c1c-b272-687221358335 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.553244] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114610, 'name': ReconfigVM_Task, 'duration_secs': 0.373327} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.556202] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Reconfigured VM instance instance-0000005a to attach disk [datastore2] b412d3e5-0536-413c-9519-5f85c7647cdf/b412d3e5-0536-413c-9519-5f85c7647cdf.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.559945] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1033.559945] env[65522]: value = "task-5114611" [ 1033.559945] env[65522]: _type = "Task" [ 1033.559945] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.562722] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8ce88083-15c5-4bce-9167-f157885741cd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.574754] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114611, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.576513] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1033.576513] env[65522]: value = "task-5114612" [ 1033.576513] env[65522]: _type = "Task" [ 1033.576513] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.589715] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114612, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.693929] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c1551d5-8da5-49c4-b321-4b901622a3a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.703851] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fd32ec8-a3f6-4a8e-bf2e-8399376da5e0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.753140] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43fdb7e4-79be-4552-b809-49fcb91d2837 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.762023] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07fbb9a9-05a4-4afb-8773-7d2ec431f440 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.776813] env[65522]: DEBUG nova.compute.provider_tree [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.926150] env[65522]: DEBUG nova.compute.manager [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1033.951844] env[65522]: DEBUG nova.virt.hardware [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1033.952505] env[65522]: DEBUG nova.virt.hardware [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1033.952505] env[65522]: DEBUG nova.virt.hardware [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1033.952505] env[65522]: DEBUG nova.virt.hardware [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1033.952709] env[65522]: DEBUG nova.virt.hardware [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1033.952709] env[65522]: DEBUG nova.virt.hardware [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1033.952918] env[65522]: DEBUG nova.virt.hardware [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1033.953032] env[65522]: DEBUG nova.virt.hardware [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1033.953202] env[65522]: DEBUG nova.virt.hardware [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1033.953361] env[65522]: DEBUG nova.virt.hardware [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1033.953526] env[65522]: DEBUG nova.virt.hardware [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1033.954426] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87743b23-5e7a-493a-8e40-acc52a82e32a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.963958] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a72ea29b-590d-42bb-be0b-de97ccc243c4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.085937] env[65522]: INFO nova.compute.manager [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Took 38.00 seconds to build instance. [ 1034.087227] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114611, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073953} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.087769] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1034.088780] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b68b15-bcea-48c1-a5a0-d31aab906933 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.094274] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114612, 'name': Rename_Task, 'duration_secs': 0.169754} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.094878] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1034.095442] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36336534-618a-4392-8631-793b96d4d036 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.115985] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 7a358133-88a5-4dd7-ab69-45b8172870a5/7a358133-88a5-4dd7-ab69-45b8172870a5.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1034.117927] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e52d006-dcde-413e-bb89-ddb2d7df126d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.133384] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1034.133384] env[65522]: value = "task-5114613" [ 1034.133384] env[65522]: _type = "Task" [ 1034.133384] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.141980] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1034.141980] env[65522]: value = "task-5114614" [ 1034.141980] env[65522]: _type = "Task" [ 1034.141980] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.149292] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114613, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.155647] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114614, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.280452] env[65522]: DEBUG nova.scheduler.client.report [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1034.589259] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9108743c-6306-48aa-8786-cecf655054f8 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "99353848-2f0f-4388-9fcd-91e799342386" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.515s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1034.644723] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114613, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.654377] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114614, 'name': ReconfigVM_Task, 'duration_secs': 0.406728} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.654663] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 7a358133-88a5-4dd7-ab69-45b8172870a5/7a358133-88a5-4dd7-ab69-45b8172870a5.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1034.655636] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-793bde34-4afc-437e-8f93-16af62224eab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.662692] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1034.662692] env[65522]: value = "task-5114615" [ 1034.662692] env[65522]: _type = "Task" [ 1034.662692] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.672023] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114615, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.939894] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.546s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1034.941802] env[65522]: DEBUG oslo_concurrency.lockutils [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.993s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1034.942261] env[65522]: DEBUG nova.objects.instance [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lazy-loading 'resources' on Instance uuid 9c39b722-56ff-44fa-8f66-3e3432645a68 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.012696] env[65522]: DEBUG nova.compute.manager [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Received event network-changed-18edb16a-a08b-4e97-9b0a-a138798ed559 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1035.014143] env[65522]: DEBUG nova.compute.manager [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Refreshing instance network info cache due to event network-changed-18edb16a-a08b-4e97-9b0a-a138798ed559. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1035.014143] env[65522]: DEBUG oslo_concurrency.lockutils [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] Acquiring lock "refresh_cache-99353848-2f0f-4388-9fcd-91e799342386" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.014143] env[65522]: DEBUG oslo_concurrency.lockutils [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] Acquired lock "refresh_cache-99353848-2f0f-4388-9fcd-91e799342386" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1035.014143] env[65522]: DEBUG nova.network.neutron [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Refreshing network info cache for port 18edb16a-a08b-4e97-9b0a-a138798ed559 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1035.147603] env[65522]: DEBUG oslo_vmware.api [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114613, 'name': PowerOnVM_Task, 'duration_secs': 0.552216} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.148438] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1035.148438] env[65522]: INFO nova.compute.manager [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Took 5.14 seconds to spawn the instance on the hypervisor. [ 1035.148563] env[65522]: DEBUG nova.compute.manager [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1035.149469] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3527eef6-1d46-46f9-9193-d368feec36a3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.174549] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114615, 'name': Rename_Task, 'duration_secs': 0.161657} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.175011] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1035.175417] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e4adb7d-53ed-4663-8aaf-c8c2c6ce3163 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.185055] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1035.185055] env[65522]: value = "task-5114616" [ 1035.185055] env[65522]: _type = "Task" [ 1035.185055] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.193994] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114616, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.212433] env[65522]: DEBUG nova.network.neutron [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Successfully updated port: d443e903-a2fa-4c64-b185-81fe54af293a {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1035.275380] env[65522]: DEBUG nova.compute.manager [req-4e855213-1f4e-4843-a314-26d6986feee4 req-4bb8893a-ad4d-4381-b8c8-54328a95831f service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Received event network-vif-plugged-d443e903-a2fa-4c64-b185-81fe54af293a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1035.275380] env[65522]: DEBUG oslo_concurrency.lockutils [req-4e855213-1f4e-4843-a314-26d6986feee4 req-4bb8893a-ad4d-4381-b8c8-54328a95831f service nova] Acquiring lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1035.275380] env[65522]: DEBUG oslo_concurrency.lockutils [req-4e855213-1f4e-4843-a314-26d6986feee4 req-4bb8893a-ad4d-4381-b8c8-54328a95831f service nova] Lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.275380] env[65522]: DEBUG oslo_concurrency.lockutils [req-4e855213-1f4e-4843-a314-26d6986feee4 req-4bb8893a-ad4d-4381-b8c8-54328a95831f service nova] Lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.275380] env[65522]: DEBUG nova.compute.manager [req-4e855213-1f4e-4843-a314-26d6986feee4 req-4bb8893a-ad4d-4381-b8c8-54328a95831f service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] No waiting events found dispatching network-vif-plugged-d443e903-a2fa-4c64-b185-81fe54af293a {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1035.275380] env[65522]: WARNING nova.compute.manager [req-4e855213-1f4e-4843-a314-26d6986feee4 req-4bb8893a-ad4d-4381-b8c8-54328a95831f service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Received unexpected event network-vif-plugged-d443e903-a2fa-4c64-b185-81fe54af293a for instance with vm_state building and task_state spawning. [ 1035.452703] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2fb39fa-e317-4cf4-a190-bc5c2a8a78ab tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 34.668s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.453048] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 8.609s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.453248] env[65522]: INFO nova.compute.manager [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Unshelving [ 1035.516344] env[65522]: WARNING neutronclient.v2_0.client [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1035.516832] env[65522]: WARNING openstack [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.517221] env[65522]: WARNING openstack [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.657853] env[65522]: WARNING openstack [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.658254] env[65522]: WARNING openstack [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.674598] env[65522]: INFO nova.compute.manager [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Took 34.15 seconds to build instance. [ 1035.698449] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114616, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.717912] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.717912] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1035.717912] env[65522]: DEBUG nova.network.neutron [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1035.750969] env[65522]: WARNING neutronclient.v2_0.client [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1035.751791] env[65522]: WARNING openstack [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1035.752564] env[65522]: WARNING openstack [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1035.762783] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2a4ac6-1a6c-4f22-a371-d79dfdff5a9d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.772277] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-983ca248-5d32-47fa-8cec-88977e31a80a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.809751] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c5dc22-55a6-4cb6-8d3c-89434678e699 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.818450] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f087a5-71bd-4f36-a526-ecdfdb434f1f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.833656] env[65522]: DEBUG nova.compute.provider_tree [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.857214] env[65522]: DEBUG nova.network.neutron [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Updated VIF entry in instance network info cache for port 18edb16a-a08b-4e97-9b0a-a138798ed559. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1035.857611] env[65522]: DEBUG nova.network.neutron [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Updating instance_info_cache with network_info: [{"id": "18edb16a-a08b-4e97-9b0a-a138798ed559", "address": "fa:16:3e:f6:b9:dd", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18edb16a-a0", "ovs_interfaceid": "18edb16a-a08b-4e97-9b0a-a138798ed559", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1036.180199] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dcf08ff7-7867-4001-9362-8825c0b0740c tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Lock "b412d3e5-0536-413c-9519-5f85c7647cdf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.663s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1036.195249] env[65522]: DEBUG oslo_vmware.api [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114616, 'name': PowerOnVM_Task, 'duration_secs': 0.545924} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.195622] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1036.195876] env[65522]: INFO nova.compute.manager [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Took 8.71 seconds to spawn the instance on the hypervisor. [ 1036.196151] env[65522]: DEBUG nova.compute.manager [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1036.197226] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9efc1c5f-1d95-41b7-8b72-b97a8ff668fc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.222783] env[65522]: WARNING openstack [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1036.223092] env[65522]: WARNING openstack [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1036.234041] env[65522]: INFO nova.compute.manager [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Rebuilding instance [ 1036.283925] env[65522]: DEBUG nova.network.neutron [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1036.298463] env[65522]: DEBUG nova.compute.manager [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1036.299586] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eff009b7-048c-4643-ab06-e5453cb79152 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.307113] env[65522]: WARNING openstack [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1036.307554] env[65522]: WARNING openstack [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1036.337277] env[65522]: DEBUG nova.scheduler.client.report [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1036.360821] env[65522]: DEBUG oslo_concurrency.lockutils [req-79054ee4-66b1-4878-b71a-6e15a1041ef2 req-96ed8070-dd7c-489e-a700-96afdcaecafe service nova] Releasing lock "refresh_cache-99353848-2f0f-4388-9fcd-91e799342386" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1036.394291] env[65522]: WARNING neutronclient.v2_0.client [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1036.394966] env[65522]: WARNING openstack [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1036.395337] env[65522]: WARNING openstack [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1036.465646] env[65522]: DEBUG nova.compute.utils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1036.495350] env[65522]: DEBUG nova.network.neutron [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updating instance_info_cache with network_info: [{"id": "d443e903-a2fa-4c64-b185-81fe54af293a", "address": "fa:16:3e:81:ed:9d", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd443e903-a2", "ovs_interfaceid": "d443e903-a2fa-4c64-b185-81fe54af293a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1036.719207] env[65522]: INFO nova.compute.manager [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Took 37.43 seconds to build instance. [ 1036.842375] env[65522]: DEBUG oslo_concurrency.lockutils [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.901s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1036.844846] env[65522]: DEBUG oslo_concurrency.lockutils [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.908s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1036.845095] env[65522]: DEBUG nova.objects.instance [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lazy-loading 'resources' on Instance uuid 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.864875] env[65522]: INFO nova.scheduler.client.report [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleted allocations for instance 9c39b722-56ff-44fa-8f66-3e3432645a68 [ 1036.969649] env[65522]: INFO nova.virt.block_device [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Booting with volume fd5bf7dd-1d57-4e86-badc-405c9e3774cc at /dev/sdb [ 1036.998921] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1036.999087] env[65522]: DEBUG nova.compute.manager [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Instance network_info: |[{"id": "d443e903-a2fa-4c64-b185-81fe54af293a", "address": "fa:16:3e:81:ed:9d", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd443e903-a2", "ovs_interfaceid": "d443e903-a2fa-4c64-b185-81fe54af293a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1036.999710] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:ed:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4345ef6-a7c8-4c1c-badf-a0d4f578b61c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd443e903-a2fa-4c64-b185-81fe54af293a', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1037.007509] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1037.008856] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1037.009127] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7147e2b8-cae0-49a1-bd72-c1d1ca3c247f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.011219] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-126fd578-672f-4c5c-a6bc-7547df4e4d3d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.037348] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f16f3d-2c3d-4d50-8f87-9e6e131fc7b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.048454] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1037.048454] env[65522]: value = "task-5114617" [ 1037.048454] env[65522]: _type = "Task" [ 1037.048454] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.061362] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114617, 'name': CreateVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.074043] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-16d412ac-9306-491b-9cfa-ddffaa4c1281 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.084425] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fff769a-b9bb-423a-a559-d095c2bd5444 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.121178] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1f2700-ba9d-4b3d-9b16-4dc4aaa83054 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.131768] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e00d5f-9317-44fb-9539-06e1b191f31c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.147657] env[65522]: DEBUG nova.virt.block_device [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Updating existing volume attachment record: 980c2a76-ba44-415d-823b-0f303e04d510 {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1037.221400] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bba5d46d-d749-4ff8-b7f0-9ae3cbe75b5d tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "7a358133-88a5-4dd7-ab69-45b8172870a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.942s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1037.308253] env[65522]: DEBUG nova.compute.manager [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Received event network-changed-d443e903-a2fa-4c64-b185-81fe54af293a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1037.308253] env[65522]: DEBUG nova.compute.manager [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Refreshing instance network info cache due to event network-changed-d443e903-a2fa-4c64-b185-81fe54af293a. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1037.308253] env[65522]: DEBUG oslo_concurrency.lockutils [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] Acquiring lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.308383] env[65522]: DEBUG oslo_concurrency.lockutils [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] Acquired lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1037.308631] env[65522]: DEBUG nova.network.neutron [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Refreshing network info cache for port d443e903-a2fa-4c64-b185-81fe54af293a {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1037.324032] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1037.325118] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1558b3c6-4f3d-4657-b43b-074be84cc8cc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.333861] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1037.333861] env[65522]: value = "task-5114619" [ 1037.333861] env[65522]: _type = "Task" [ 1037.333861] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.345215] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.377309] env[65522]: DEBUG oslo_concurrency.lockutils [None req-32e03632-366b-4e30-bf69-c2ed947ec151 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "9c39b722-56ff-44fa-8f66-3e3432645a68" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.674s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1037.565598] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114617, 'name': CreateVM_Task, 'duration_secs': 0.456653} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.565598] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1037.565598] env[65522]: WARNING neutronclient.v2_0.client [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1037.566089] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.566089] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1037.567193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1037.567418] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-230f4b94-608e-45cf-a0e5-93fae3a1683a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.575290] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1037.575290] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5211f5cb-9fdf-5c10-c43e-5f6b4e2012e9" [ 1037.575290] env[65522]: _type = "Task" [ 1037.575290] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.587274] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5211f5cb-9fdf-5c10-c43e-5f6b4e2012e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.633528] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8546fc07-6f3d-479d-a0fc-6d7e66f77148 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.644667] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7163b14b-9934-4a61-8fbd-70999e6f0db7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.696874] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d8696e4-13a7-4a12-ad7f-1e58f1526fc0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.703586] env[65522]: DEBUG oslo_concurrency.lockutils [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "7a358133-88a5-4dd7-ab69-45b8172870a5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1037.703938] env[65522]: DEBUG oslo_concurrency.lockutils [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "7a358133-88a5-4dd7-ab69-45b8172870a5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1037.710674] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-340a6084-d82e-40a0-9b08-adfa287e09ca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.727589] env[65522]: DEBUG nova.compute.provider_tree [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.813967] env[65522]: WARNING neutronclient.v2_0.client [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1037.814699] env[65522]: WARNING openstack [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.815167] env[65522]: WARNING openstack [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.845823] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114619, 'name': PowerOffVM_Task, 'duration_secs': 0.165602} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.846144] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1037.847016] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1037.847943] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3416bad9-b0c6-4f76-922c-73c511cae544 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.855916] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1037.858556] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-831e0eac-16f9-4baa-8523-fa09a5405df0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.889661] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1037.890034] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1037.890310] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Deleting the datastore file [datastore2] b412d3e5-0536-413c-9519-5f85c7647cdf {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1037.890721] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f1d7211-50c8-49ce-91bf-e4ce67ed7580 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.899477] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1037.899477] env[65522]: value = "task-5114623" [ 1037.899477] env[65522]: _type = "Task" [ 1037.899477] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.918567] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114623, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.924403] env[65522]: WARNING openstack [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.924864] env[65522]: WARNING openstack [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.993054] env[65522]: WARNING neutronclient.v2_0.client [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1037.993554] env[65522]: WARNING openstack [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.994067] env[65522]: WARNING openstack [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1038.089802] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5211f5cb-9fdf-5c10-c43e-5f6b4e2012e9, 'name': SearchDatastore_Task, 'duration_secs': 0.015578} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.092374] env[65522]: DEBUG nova.network.neutron [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updated VIF entry in instance network info cache for port d443e903-a2fa-4c64-b185-81fe54af293a. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1038.092848] env[65522]: DEBUG nova.network.neutron [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updating instance_info_cache with network_info: [{"id": "d443e903-a2fa-4c64-b185-81fe54af293a", "address": "fa:16:3e:81:ed:9d", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd443e903-a2", "ovs_interfaceid": "d443e903-a2fa-4c64-b185-81fe54af293a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1038.094294] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1038.094546] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1038.094872] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.095033] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1038.095231] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1038.095855] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15a07d68-484a-4583-bf04-b835b75cd769 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.106549] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1038.106753] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1038.107952] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9f22a02-8bc3-4443-81a0-6e12fbce6d1c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.114536] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1038.114536] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d0d91b-f761-296c-a9c9-cba278515b35" [ 1038.114536] env[65522]: _type = "Task" [ 1038.114536] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.123331] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d0d91b-f761-296c-a9c9-cba278515b35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.211994] env[65522]: DEBUG nova.compute.utils [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1038.230837] env[65522]: DEBUG nova.scheduler.client.report [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1038.417041] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114623, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164922} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.417041] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1038.417387] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1038.417387] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1038.596991] env[65522]: DEBUG oslo_concurrency.lockutils [req-5aae5b78-90d4-478a-846e-66c81d112124 req-9eea8dda-1c6b-4453-96bd-4d8c625a986e service nova] Releasing lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1038.625560] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d0d91b-f761-296c-a9c9-cba278515b35, 'name': SearchDatastore_Task, 'duration_secs': 0.011843} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.626424] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24527f49-0d33-434b-97c1-18be385e8a35 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.633046] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1038.633046] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52903bc1-a50e-3c02-04bb-95b680291fd1" [ 1038.633046] env[65522]: _type = "Task" [ 1038.633046] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.643253] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52903bc1-a50e-3c02-04bb-95b680291fd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.714237] env[65522]: DEBUG oslo_concurrency.lockutils [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "7a358133-88a5-4dd7-ab69-45b8172870a5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1038.738524] env[65522]: DEBUG oslo_concurrency.lockutils [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.894s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1038.741110] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.294s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1038.741375] env[65522]: DEBUG nova.objects.instance [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lazy-loading 'resources' on Instance uuid 95a10a6f-8c6c-435d-a360-dff10eee0855 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.757616] env[65522]: INFO nova.scheduler.client.report [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Deleted allocations for instance 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7 [ 1039.153667] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52903bc1-a50e-3c02-04bb-95b680291fd1, 'name': SearchDatastore_Task, 'duration_secs': 0.012269} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.153667] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1039.154100] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 95f0772d-6d3e-4e37-a0d2-9ab9de63d124/95f0772d-6d3e-4e37-a0d2-9ab9de63d124.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1039.154386] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f4712a4-822e-4831-b62c-bbc9d48a6146 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.163513] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1039.163513] env[65522]: value = "task-5114624" [ 1039.163513] env[65522]: _type = "Task" [ 1039.163513] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.174915] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114624, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.266917] env[65522]: DEBUG oslo_concurrency.lockutils [None req-287faa17-58a1-4032-b944-295911b0b038 tempest-ServerDiskConfigTestJSON-1984357202 tempest-ServerDiskConfigTestJSON-1984357202-project-member] Lock "5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.486s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1039.463818] env[65522]: DEBUG nova.virt.hardware [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1039.464095] env[65522]: DEBUG nova.virt.hardware [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1039.464242] env[65522]: DEBUG nova.virt.hardware [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1039.464429] env[65522]: DEBUG nova.virt.hardware [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1039.464571] env[65522]: DEBUG nova.virt.hardware [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1039.464716] env[65522]: DEBUG nova.virt.hardware [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1039.464927] env[65522]: DEBUG nova.virt.hardware [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1039.465102] env[65522]: DEBUG nova.virt.hardware [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1039.465276] env[65522]: DEBUG nova.virt.hardware [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1039.465439] env[65522]: DEBUG nova.virt.hardware [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1039.465612] env[65522]: DEBUG nova.virt.hardware [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1039.466562] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f6e71a-f821-443d-b4f6-b5444ca30136 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.483179] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4924a6f8-5ce8-435f-870c-61e303228063 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.503351] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1039.510462] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1039.516026] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1039.516026] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fd45a876-05f7-4c2e-8c62-cc01494e60dd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.528732] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c210ba-f161-4b0d-8461-e2c36805306f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.541335] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1039.541335] env[65522]: value = "task-5114625" [ 1039.541335] env[65522]: _type = "Task" [ 1039.541335] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.542661] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf99c9f8-b0c1-4565-9833-2ecdca237b43 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.584797] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8e3406-c25a-406c-aab4-ad9647eefa98 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.588020] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114625, 'name': CreateVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.597184] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa17cda-6a46-4ef3-8031-dee60c894e8e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.616882] env[65522]: DEBUG nova.compute.provider_tree [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.660346] env[65522]: DEBUG nova.compute.manager [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Stashing vm_state: active {{(pid=65522) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1039.674775] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114624, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.757788] env[65522]: DEBUG oslo_concurrency.lockutils [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "7a358133-88a5-4dd7-ab69-45b8172870a5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1039.758111] env[65522]: DEBUG oslo_concurrency.lockutils [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "7a358133-88a5-4dd7-ab69-45b8172870a5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1039.758428] env[65522]: INFO nova.compute.manager [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Attaching volume 77787cc2-a082-403d-aca0-bc1b4253df4e to /dev/sdb [ 1039.802219] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0771940-0eab-4003-82a0-6bc9e22d8c93 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.812500] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c83bbe00-461f-4017-b7e7-609ef95f2233 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.829026] env[65522]: DEBUG nova.virt.block_device [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Updating existing volume attachment record: 6a078c90-8cd0-4752-8aba-9ba6143d4ed0 {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1040.050659] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114625, 'name': CreateVM_Task, 'duration_secs': 0.425597} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.050849] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1040.051383] env[65522]: DEBUG oslo_concurrency.lockutils [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.051611] env[65522]: DEBUG oslo_concurrency.lockutils [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1040.051966] env[65522]: DEBUG oslo_concurrency.lockutils [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1040.052261] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d81c5ce3-fa6a-49a7-be7c-9738267d10d4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.057337] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1040.057337] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5278e126-56ec-9097-2a4b-af7732d0b55a" [ 1040.057337] env[65522]: _type = "Task" [ 1040.057337] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.065941] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5278e126-56ec-9097-2a4b-af7732d0b55a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.120699] env[65522]: DEBUG nova.scheduler.client.report [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1040.178641] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114624, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513734} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.178983] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 95f0772d-6d3e-4e37-a0d2-9ab9de63d124/95f0772d-6d3e-4e37-a0d2-9ab9de63d124.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1040.179288] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1040.179498] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3307bd7a-1137-48be-9e2f-6712b384cf6e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.182588] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1040.187321] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1040.187321] env[65522]: value = "task-5114628" [ 1040.187321] env[65522]: _type = "Task" [ 1040.187321] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.196393] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114628, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.571517] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5278e126-56ec-9097-2a4b-af7732d0b55a, 'name': SearchDatastore_Task, 'duration_secs': 0.010691} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.571838] env[65522]: DEBUG oslo_concurrency.lockutils [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1040.572092] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1040.572356] env[65522]: DEBUG oslo_concurrency.lockutils [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.572499] env[65522]: DEBUG oslo_concurrency.lockutils [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1040.572672] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1040.572958] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c88beb5c-f86f-4992-a846-2c4d20c8f5e9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.582895] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1040.583098] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1040.584131] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95bc3cda-2e12-4864-9f5e-72da795e7e2a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.589923] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1040.589923] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a8393c-839f-6fdb-02cc-2a942c028850" [ 1040.589923] env[65522]: _type = "Task" [ 1040.589923] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.598625] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a8393c-839f-6fdb-02cc-2a942c028850, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.628120] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.887s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1040.631310] env[65522]: DEBUG oslo_concurrency.lockutils [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.918s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1040.631310] env[65522]: DEBUG nova.objects.instance [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lazy-loading 'resources' on Instance uuid e7baf2a0-21dd-4610-8230-81b99b64856b {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.652226] env[65522]: INFO nova.scheduler.client.report [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Deleted allocations for instance 95a10a6f-8c6c-435d-a360-dff10eee0855 [ 1040.699806] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114628, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.116348} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.700122] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1040.700984] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d027829-b7e8-4873-9174-6a1de4d42a89 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.727955] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 95f0772d-6d3e-4e37-a0d2-9ab9de63d124/95f0772d-6d3e-4e37-a0d2-9ab9de63d124.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.728996] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ab12534-5dff-49ff-93c2-110c8760e13f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.752192] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1040.752192] env[65522]: value = "task-5114631" [ 1040.752192] env[65522]: _type = "Task" [ 1040.752192] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.762443] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114631, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.101573] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a8393c-839f-6fdb-02cc-2a942c028850, 'name': SearchDatastore_Task, 'duration_secs': 0.011638} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.102360] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f2176b6-7661-47f9-bb34-a6893e6c996a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.116710] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1041.116710] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526d695c-ab05-48a5-29d4-441345ae39a9" [ 1041.116710] env[65522]: _type = "Task" [ 1041.116710] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.126530] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526d695c-ab05-48a5-29d4-441345ae39a9, 'name': SearchDatastore_Task, 'duration_secs': 0.011137} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.126881] env[65522]: DEBUG oslo_concurrency.lockutils [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1041.127204] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] b412d3e5-0536-413c-9519-5f85c7647cdf/b412d3e5-0536-413c-9519-5f85c7647cdf.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1041.127739] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da3329c1-3146-44b2-8cd5-2dbef3c7e773 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.139424] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1041.139424] env[65522]: value = "task-5114632" [ 1041.139424] env[65522]: _type = "Task" [ 1041.139424] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.149613] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114632, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.160922] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5a8386ac-f4f9-4fc9-b2cc-fb737f9a8c62 tempest-ServerShowV247Test-1510773036 tempest-ServerShowV247Test-1510773036-project-member] Lock "95a10a6f-8c6c-435d-a360-dff10eee0855" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.479s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1041.263500] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114631, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.444861] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca7536a-650b-4376-9b81-a4a0ea205ddc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.457374] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c44bc21-bca3-4daf-bc48-1fc3b4daa136 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.491904] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f295cd8-6a0f-443c-9590-a2608a01d5a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.504404] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ff6645-324e-493d-bfc1-8ad040e1940c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.528042] env[65522]: DEBUG nova.compute.provider_tree [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.652536] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114632, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.763252] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114631, 'name': ReconfigVM_Task, 'duration_secs': 0.67314} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.763609] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 95f0772d-6d3e-4e37-a0d2-9ab9de63d124/95f0772d-6d3e-4e37-a0d2-9ab9de63d124.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1041.764316] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fe77b4f2-ec98-4c2b-94a2-657beb31d853 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.772521] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1041.772521] env[65522]: value = "task-5114633" [ 1041.772521] env[65522]: _type = "Task" [ 1041.772521] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.783049] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114633, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.031969] env[65522]: DEBUG nova.scheduler.client.report [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1042.151429] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114632, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587855} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.151700] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] b412d3e5-0536-413c-9519-5f85c7647cdf/b412d3e5-0536-413c-9519-5f85c7647cdf.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1042.151919] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1042.152242] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-386e3b65-d1d4-427f-a012-d9f493a2f6de {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.160124] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1042.160124] env[65522]: value = "task-5114634" [ 1042.160124] env[65522]: _type = "Task" [ 1042.160124] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.169572] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114634, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.283503] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114633, 'name': Rename_Task, 'duration_secs': 0.161015} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.283728] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1042.283993] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4cad63ac-56d7-414f-9813-e277f9191838 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.292266] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1042.292266] env[65522]: value = "task-5114635" [ 1042.292266] env[65522]: _type = "Task" [ 1042.292266] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.302332] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114635, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.305829] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1042.306209] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1042.536896] env[65522]: DEBUG oslo_concurrency.lockutils [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.906s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1042.539414] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.035s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1042.541564] env[65522]: INFO nova.compute.claims [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1042.563973] env[65522]: INFO nova.scheduler.client.report [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Deleted allocations for instance e7baf2a0-21dd-4610-8230-81b99b64856b [ 1042.671096] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114634, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07395} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.671451] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1042.672316] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a70e8a4-1abc-452e-be89-f0dccec6b085 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.694553] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] b412d3e5-0536-413c-9519-5f85c7647cdf/b412d3e5-0536-413c-9519-5f85c7647cdf.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1042.694863] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9580ff74-1b2e-4a69-87c6-a171803a2b0f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.717424] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1042.717424] env[65522]: value = "task-5114637" [ 1042.717424] env[65522]: _type = "Task" [ 1042.717424] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.726583] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114637, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.803097] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114635, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.810371] env[65522]: DEBUG nova.compute.utils [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1043.072549] env[65522]: DEBUG oslo_concurrency.lockutils [None req-daba28d6-6b7f-41aa-b81f-779dd7a553ea tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "e7baf2a0-21dd-4610-8230-81b99b64856b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.621s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1043.230400] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114637, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.268667] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1043.304208] env[65522]: DEBUG oslo_vmware.api [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114635, 'name': PowerOnVM_Task, 'duration_secs': 0.965787} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.308903] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1043.308903] env[65522]: INFO nova.compute.manager [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Took 9.38 seconds to spawn the instance on the hypervisor. [ 1043.309123] env[65522]: DEBUG nova.compute.manager [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1043.310399] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14fdcd63-a519-4b53-aad5-7748d507ce8c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.313594] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1043.729889] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114637, 'name': ReconfigVM_Task, 'duration_secs': 0.825114} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.735349] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Reconfigured VM instance instance-0000005a to attach disk [datastore2] b412d3e5-0536-413c-9519-5f85c7647cdf/b412d3e5-0536-413c-9519-5f85c7647cdf.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1043.736835] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce54dd1f-3099-462a-8d92-d98821964f97 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.745201] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1043.745201] env[65522]: value = "task-5114638" [ 1043.745201] env[65522]: _type = "Task" [ 1043.745201] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.756434] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114638, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.803886] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-149e40ba-a089-40a5-b19b-7c16e2787132 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.816464] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-782ab8c4-0a43-4d65-bca9-31622b18bb1b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.864170] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a93e4fe4-c054-42c3-8199-6a5ff4d0382a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.867761] env[65522]: INFO nova.compute.manager [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Took 23.08 seconds to build instance. [ 1043.875753] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e0ed5f-e9f7-4cb9-8306-4d848d3b3777 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.895702] env[65522]: DEBUG nova.compute.provider_tree [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.257770] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114638, 'name': Rename_Task, 'duration_secs': 0.190733} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.257770] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1044.257770] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-939fc160-25a9-41fc-a71c-3934b3a7da13 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.267212] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1044.267212] env[65522]: value = "task-5114639" [ 1044.267212] env[65522]: _type = "Task" [ 1044.267212] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.277824] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114639, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.370187] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7843b24a-599d-445e-ac44-86113382b5ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.594s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1044.399738] env[65522]: DEBUG nova.scheduler.client.report [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1044.421778] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1044.422122] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1044.424049] env[65522]: INFO nova.compute.manager [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Attaching volume a517fee4-4f9a-40e5-9c5c-0240aebef0cc to /dev/sdb [ 1044.464227] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1bf1c13-6cad-4771-a2e0-799b866ff2e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.475910] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf892e3-3de1-4fba-92d3-c3dac0e02a19 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.492263] env[65522]: DEBUG nova.virt.block_device [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Updating existing volume attachment record: 33fba941-86d3-44f1-92fc-c43c88fb4db3 {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1044.790083] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114639, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.884156] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1044.884510] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994904', 'volume_id': '77787cc2-a082-403d-aca0-bc1b4253df4e', 'name': 'volume-77787cc2-a082-403d-aca0-bc1b4253df4e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '7a358133-88a5-4dd7-ab69-45b8172870a5', 'attached_at': '', 'detached_at': '', 'volume_id': '77787cc2-a082-403d-aca0-bc1b4253df4e', 'serial': '77787cc2-a082-403d-aca0-bc1b4253df4e'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1044.886637] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e899d503-6e73-4978-95d3-8c7abd051843 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.912529] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1044.912529] env[65522]: DEBUG nova.compute.manager [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1044.915930] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.950s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1044.916199] env[65522]: DEBUG nova.objects.instance [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lazy-loading 'resources' on Instance uuid 43691901-7ac0-4523-a19c-eedf5ae4c040 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.918352] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf4c0b7-3854-4ae2-9127-847a770a2b9e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.954904] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] volume-77787cc2-a082-403d-aca0-bc1b4253df4e/volume-77787cc2-a082-403d-aca0-bc1b4253df4e.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1044.957156] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e6d0802-97c7-4e22-8810-9a75aa9110c9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.986356] env[65522]: DEBUG oslo_vmware.api [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1044.986356] env[65522]: value = "task-5114642" [ 1044.986356] env[65522]: _type = "Task" [ 1044.986356] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.000746] env[65522]: DEBUG oslo_vmware.api [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114642, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.282099] env[65522]: DEBUG oslo_vmware.api [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114639, 'name': PowerOnVM_Task, 'duration_secs': 0.898915} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.282758] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1045.284019] env[65522]: DEBUG nova.compute.manager [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1045.284368] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-988293d3-3bea-4a8a-80e7-1022bd9c905b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.418097] env[65522]: DEBUG nova.compute.utils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1045.428726] env[65522]: DEBUG nova.compute.manager [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1045.429444] env[65522]: DEBUG nova.network.neutron [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1045.429581] env[65522]: WARNING neutronclient.v2_0.client [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1045.429865] env[65522]: WARNING neutronclient.v2_0.client [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1045.430564] env[65522]: WARNING openstack [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1045.430834] env[65522]: WARNING openstack [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1045.505029] env[65522]: DEBUG oslo_vmware.api [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114642, 'name': ReconfigVM_Task, 'duration_secs': 0.493295} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.505029] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Reconfigured VM instance instance-00000059 to attach disk [datastore1] volume-77787cc2-a082-403d-aca0-bc1b4253df4e/volume-77787cc2-a082-403d-aca0-bc1b4253df4e.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1045.515279] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b29219de-a965-41da-8737-087b2c6561b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.534767] env[65522]: DEBUG oslo_vmware.api [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1045.534767] env[65522]: value = "task-5114644" [ 1045.534767] env[65522]: _type = "Task" [ 1045.534767] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.548729] env[65522]: DEBUG oslo_vmware.api [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114644, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.759882] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad48599-c9c0-4510-bf4d-7b55cf4a6647 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.769198] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5588ff-63bc-45be-b4f8-7da5aac13146 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.812134] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35450749-bf80-40c1-977f-5af8143b1a1f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.815721] env[65522]: DEBUG oslo_concurrency.lockutils [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1045.823154] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008c69c9-e686-408a-b209-4a33b55a2e39 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.841891] env[65522]: DEBUG nova.compute.provider_tree [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.881767] env[65522]: DEBUG nova.policy [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f51006be1b2e401a8dd5504e0e481810', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ea0fab7de6c47eeaf3bdde413ae3901', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1045.938979] env[65522]: DEBUG nova.compute.manager [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1046.048154] env[65522]: DEBUG oslo_vmware.api [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114644, 'name': ReconfigVM_Task, 'duration_secs': 0.214581} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.048496] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994904', 'volume_id': '77787cc2-a082-403d-aca0-bc1b4253df4e', 'name': 'volume-77787cc2-a082-403d-aca0-bc1b4253df4e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '7a358133-88a5-4dd7-ab69-45b8172870a5', 'attached_at': '', 'detached_at': '', 'volume_id': '77787cc2-a082-403d-aca0-bc1b4253df4e', 'serial': '77787cc2-a082-403d-aca0-bc1b4253df4e'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1046.306372] env[65522]: DEBUG nova.network.neutron [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Successfully created port: 76e82556-fba8-4c08-aa9b-1b60d15660c1 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1046.345498] env[65522]: DEBUG nova.scheduler.client.report [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1046.807996] env[65522]: DEBUG nova.compute.manager [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Received event network-changed-2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1046.807996] env[65522]: DEBUG nova.compute.manager [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Refreshing instance network info cache due to event network-changed-2b2de45d-4c08-4bb4-810a-2f26266cbc20. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1046.808487] env[65522]: DEBUG oslo_concurrency.lockutils [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] Acquiring lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.808576] env[65522]: DEBUG oslo_concurrency.lockutils [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] Acquired lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1046.808765] env[65522]: DEBUG nova.network.neutron [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Refreshing network info cache for port 2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1046.852710] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.936s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1046.855774] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.673s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1046.888522] env[65522]: INFO nova.scheduler.client.report [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleted allocations for instance 43691901-7ac0-4523-a19c-eedf5ae4c040 [ 1046.950462] env[65522]: DEBUG nova.compute.manager [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1046.987511] env[65522]: DEBUG nova.virt.hardware [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c73b4053ddd71f985d752bb490c0900a',container_format='bare',created_at=2025-12-12T19:05:14Z,direct_url=,disk_format='vmdk',id=e85cd8c4-7ac5-4359-bcda-17d18e3e60f9,min_disk=1,min_ram=0,name='tempest-test-snap-566917532',owner='3ea0fab7de6c47eeaf3bdde413ae3901',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2025-12-12T19:05:33Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1046.987728] env[65522]: DEBUG nova.virt.hardware [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1046.987925] env[65522]: DEBUG nova.virt.hardware [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1046.988318] env[65522]: DEBUG nova.virt.hardware [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1046.988449] env[65522]: DEBUG nova.virt.hardware [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1046.988669] env[65522]: DEBUG nova.virt.hardware [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1046.988947] env[65522]: DEBUG nova.virt.hardware [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1046.989242] env[65522]: DEBUG nova.virt.hardware [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1046.989469] env[65522]: DEBUG nova.virt.hardware [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1046.989742] env[65522]: DEBUG nova.virt.hardware [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1046.989853] env[65522]: DEBUG nova.virt.hardware [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1046.991169] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7d764f-8ebc-4608-bc39-fc10e597d295 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.000764] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1ad939-632f-4a4d-9a38-3ee4e6c9da9c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.110647] env[65522]: DEBUG nova.objects.instance [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lazy-loading 'flavor' on Instance uuid 7a358133-88a5-4dd7-ab69-45b8172870a5 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.314022] env[65522]: WARNING neutronclient.v2_0.client [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1047.314022] env[65522]: WARNING openstack [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1047.314022] env[65522]: WARNING openstack [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1047.363619] env[65522]: INFO nova.compute.claims [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1047.472052] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fece5e11-f000-4f5b-b984-71ba478309ec tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "43691901-7ac0-4523-a19c-eedf5ae4c040" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.013s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1047.614016] env[65522]: DEBUG oslo_concurrency.lockutils [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquiring lock "b412d3e5-0536-413c-9519-5f85c7647cdf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1047.614243] env[65522]: DEBUG oslo_concurrency.lockutils [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Lock "b412d3e5-0536-413c-9519-5f85c7647cdf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1047.614464] env[65522]: DEBUG oslo_concurrency.lockutils [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquiring lock "b412d3e5-0536-413c-9519-5f85c7647cdf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1047.614919] env[65522]: DEBUG oslo_concurrency.lockutils [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Lock "b412d3e5-0536-413c-9519-5f85c7647cdf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1047.615605] env[65522]: DEBUG oslo_concurrency.lockutils [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Lock "b412d3e5-0536-413c-9519-5f85c7647cdf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1047.619824] env[65522]: INFO nova.compute.manager [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Terminating instance [ 1047.621914] env[65522]: DEBUG oslo_concurrency.lockutils [None req-beb00118-0af7-40ab-9748-446a7c65f417 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "7a358133-88a5-4dd7-ab69-45b8172870a5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.864s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1047.650234] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1047.650468] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1047.975168] env[65522]: INFO nova.compute.resource_tracker [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating resource usage from migration 5c0d4fa1-1069-4eb5-8c35-44a984e954e4 [ 1048.056850] env[65522]: DEBUG nova.network.neutron [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Successfully updated port: 76e82556-fba8-4c08-aa9b-1b60d15660c1 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1048.126733] env[65522]: DEBUG oslo_concurrency.lockutils [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquiring lock "refresh_cache-b412d3e5-0536-413c-9519-5f85c7647cdf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.127108] env[65522]: DEBUG oslo_concurrency.lockutils [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquired lock "refresh_cache-b412d3e5-0536-413c-9519-5f85c7647cdf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1048.127108] env[65522]: DEBUG nova.network.neutron [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1048.154262] env[65522]: DEBUG nova.compute.manager [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1048.222188] env[65522]: DEBUG nova.compute.manager [req-94e0889b-f8c4-4321-b300-4e6080d3df5e req-7a839230-22d3-4dab-a845-bb0676e8c2f7 service nova] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Received event network-vif-plugged-76e82556-fba8-4c08-aa9b-1b60d15660c1 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1048.222486] env[65522]: DEBUG oslo_concurrency.lockutils [req-94e0889b-f8c4-4321-b300-4e6080d3df5e req-7a839230-22d3-4dab-a845-bb0676e8c2f7 service nova] Acquiring lock "f1ce2d19-347a-4830-8501-8cceb87d4bf8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.222626] env[65522]: DEBUG oslo_concurrency.lockutils [req-94e0889b-f8c4-4321-b300-4e6080d3df5e req-7a839230-22d3-4dab-a845-bb0676e8c2f7 service nova] Lock "f1ce2d19-347a-4830-8501-8cceb87d4bf8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.222790] env[65522]: DEBUG oslo_concurrency.lockutils [req-94e0889b-f8c4-4321-b300-4e6080d3df5e req-7a839230-22d3-4dab-a845-bb0676e8c2f7 service nova] Lock "f1ce2d19-347a-4830-8501-8cceb87d4bf8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.222955] env[65522]: DEBUG nova.compute.manager [req-94e0889b-f8c4-4321-b300-4e6080d3df5e req-7a839230-22d3-4dab-a845-bb0676e8c2f7 service nova] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] No waiting events found dispatching network-vif-plugged-76e82556-fba8-4c08-aa9b-1b60d15660c1 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1048.223133] env[65522]: WARNING nova.compute.manager [req-94e0889b-f8c4-4321-b300-4e6080d3df5e req-7a839230-22d3-4dab-a845-bb0676e8c2f7 service nova] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Received unexpected event network-vif-plugged-76e82556-fba8-4c08-aa9b-1b60d15660c1 for instance with vm_state building and task_state spawning. [ 1048.253583] env[65522]: WARNING openstack [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1048.253971] env[65522]: WARNING openstack [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1048.262835] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc91c45-0dc4-4d3c-ab01-7ecad75b5e20 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.272443] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2db772-8e84-452e-9c7e-f69d7874a81d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.312172] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e80786-5f56-41c6-b476-308f4931829e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.320989] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912bdf50-7e09-42a7-aede-80bc554c4a66 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.336614] env[65522]: DEBUG nova.compute.provider_tree [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.347426] env[65522]: WARNING neutronclient.v2_0.client [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1048.348218] env[65522]: WARNING openstack [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1048.348583] env[65522]: WARNING openstack [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1048.447808] env[65522]: DEBUG nova.network.neutron [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updated VIF entry in instance network info cache for port 2b2de45d-4c08-4bb4-810a-2f26266cbc20. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1048.448256] env[65522]: DEBUG nova.network.neutron [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updating instance_info_cache with network_info: [{"id": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "address": "fa:16:3e:e9:bd:f6", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b2de45d-4c", "ovs_interfaceid": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1048.563853] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "refresh_cache-f1ce2d19-347a-4830-8501-8cceb87d4bf8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.564116] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "refresh_cache-f1ce2d19-347a-4830-8501-8cceb87d4bf8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1048.564403] env[65522]: DEBUG nova.network.neutron [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1048.629765] env[65522]: WARNING neutronclient.v2_0.client [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1048.630509] env[65522]: WARNING openstack [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1048.630853] env[65522]: WARNING openstack [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1048.654864] env[65522]: DEBUG nova.network.neutron [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1048.674531] env[65522]: DEBUG oslo_concurrency.lockutils [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "7a358133-88a5-4dd7-ab69-45b8172870a5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.674531] env[65522]: DEBUG oslo_concurrency.lockutils [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "7a358133-88a5-4dd7-ab69-45b8172870a5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.674714] env[65522]: DEBUG oslo_concurrency.lockutils [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "7a358133-88a5-4dd7-ab69-45b8172870a5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.674903] env[65522]: DEBUG oslo_concurrency.lockutils [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "7a358133-88a5-4dd7-ab69-45b8172870a5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.674960] env[65522]: DEBUG oslo_concurrency.lockutils [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "7a358133-88a5-4dd7-ab69-45b8172870a5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.677654] env[65522]: INFO nova.compute.manager [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Terminating instance [ 1048.681943] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.712721] env[65522]: DEBUG nova.network.neutron [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1048.841306] env[65522]: DEBUG nova.scheduler.client.report [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1048.848882] env[65522]: DEBUG nova.compute.manager [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Received event network-changed-d443e903-a2fa-4c64-b185-81fe54af293a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1048.848882] env[65522]: DEBUG nova.compute.manager [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Refreshing instance network info cache due to event network-changed-d443e903-a2fa-4c64-b185-81fe54af293a. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1048.849120] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Acquiring lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.849284] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Acquired lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1048.850267] env[65522]: DEBUG nova.network.neutron [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Refreshing network info cache for port d443e903-a2fa-4c64-b185-81fe54af293a {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1048.951445] env[65522]: DEBUG oslo_concurrency.lockutils [req-96c36e6c-dd75-4f7a-a598-a6a50d234a17 req-c60afdeb-9a2a-4288-a59d-103adfb84e30 service nova] Releasing lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.071053] env[65522]: WARNING openstack [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.071053] env[65522]: WARNING openstack [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.110473] env[65522]: DEBUG nova.network.neutron [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1049.136925] env[65522]: WARNING openstack [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.137338] env[65522]: WARNING openstack [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.183663] env[65522]: DEBUG nova.compute.manager [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1049.183663] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1049.183819] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-053427d5-da7d-4595-8496-103c4bc180a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.192772] env[65522]: DEBUG oslo_vmware.api [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1049.192772] env[65522]: value = "task-5114649" [ 1049.192772] env[65522]: _type = "Task" [ 1049.192772] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.202388] env[65522]: DEBUG oslo_vmware.api [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114649, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.211837] env[65522]: WARNING neutronclient.v2_0.client [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1049.212669] env[65522]: WARNING openstack [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.213038] env[65522]: WARNING openstack [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.220922] env[65522]: DEBUG oslo_concurrency.lockutils [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Releasing lock "refresh_cache-b412d3e5-0536-413c-9519-5f85c7647cdf" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.221400] env[65522]: DEBUG nova.compute.manager [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1049.221593] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1049.222878] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f974038-2e1e-4591-af0a-d3cb83be3f0a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.232727] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1049.233454] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba096478-15d0-4ece-87b7-6157b7282b17 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.240413] env[65522]: DEBUG oslo_vmware.api [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1049.240413] env[65522]: value = "task-5114650" [ 1049.240413] env[65522]: _type = "Task" [ 1049.240413] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.251180] env[65522]: DEBUG oslo_vmware.api [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114650, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.326361] env[65522]: DEBUG nova.network.neutron [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Updating instance_info_cache with network_info: [{"id": "76e82556-fba8-4c08-aa9b-1b60d15660c1", "address": "fa:16:3e:07:04:53", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76e82556-fb", "ovs_interfaceid": "76e82556-fba8-4c08-aa9b-1b60d15660c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1049.346945] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.491s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1049.347181] env[65522]: INFO nova.compute.manager [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Migrating [ 1049.354946] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.086s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1049.355661] env[65522]: DEBUG nova.objects.instance [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lazy-loading 'pci_requests' on Instance uuid 79dd37b5-6b30-48ab-9f00-78214cbd132d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.357556] env[65522]: WARNING neutronclient.v2_0.client [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1049.358268] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.358629] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.472906] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "210184da-48db-4926-ae00-7c0f29622810" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1049.473391] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "210184da-48db-4926-ae00-7c0f29622810" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1049.512372] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.512840] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.567555] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1049.568644] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994906', 'volume_id': 'a517fee4-4f9a-40e5-9c5c-0240aebef0cc', 'name': 'volume-a517fee4-4f9a-40e5-9c5c-0240aebef0cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5ce4a286-efd1-4bbc-a23b-931c6701cfe4', 'attached_at': '', 'detached_at': '', 'volume_id': 'a517fee4-4f9a-40e5-9c5c-0240aebef0cc', 'serial': 'a517fee4-4f9a-40e5-9c5c-0240aebef0cc'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1049.569657] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8baf2e4b-2490-4e59-8e20-a012defe345a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.593237] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c839d90-52ac-4eea-830f-1250f0a420dc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.600428] env[65522]: WARNING neutronclient.v2_0.client [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1049.601102] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.601493] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.651956] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] volume-a517fee4-4f9a-40e5-9c5c-0240aebef0cc/volume-a517fee4-4f9a-40e5-9c5c-0240aebef0cc.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.652461] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccbb03f3-60c3-43d8-8da1-ac653493a688 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.688117] env[65522]: DEBUG oslo_vmware.api [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1049.688117] env[65522]: value = "task-5114651" [ 1049.688117] env[65522]: _type = "Task" [ 1049.688117] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.700566] env[65522]: DEBUG oslo_vmware.api [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114651, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.707669] env[65522]: DEBUG oslo_vmware.api [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114649, 'name': PowerOffVM_Task, 'duration_secs': 0.252125} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.708077] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1049.708318] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1049.708546] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994904', 'volume_id': '77787cc2-a082-403d-aca0-bc1b4253df4e', 'name': 'volume-77787cc2-a082-403d-aca0-bc1b4253df4e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '7a358133-88a5-4dd7-ab69-45b8172870a5', 'attached_at': '', 'detached_at': '', 'volume_id': '77787cc2-a082-403d-aca0-bc1b4253df4e', 'serial': '77787cc2-a082-403d-aca0-bc1b4253df4e'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1049.709565] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace3abfb-0548-4b5c-9ba7-31f4c5282ca2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.736570] env[65522]: DEBUG nova.network.neutron [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updated VIF entry in instance network info cache for port d443e903-a2fa-4c64-b185-81fe54af293a. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1049.737020] env[65522]: DEBUG nova.network.neutron [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updating instance_info_cache with network_info: [{"id": "d443e903-a2fa-4c64-b185-81fe54af293a", "address": "fa:16:3e:81:ed:9d", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd443e903-a2", "ovs_interfaceid": "d443e903-a2fa-4c64-b185-81fe54af293a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1049.739835] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4defe1b-e397-401b-81e0-ea8e995be3db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.752344] env[65522]: DEBUG oslo_vmware.api [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114650, 'name': PowerOffVM_Task, 'duration_secs': 0.278461} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.755058] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1049.755289] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1049.755787] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-64f6e412-d1ad-4473-8078-86ce05a331d6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.757758] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47bd73c8-d5a6-4cd2-9dab-611db9ed9149 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.784483] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e44a21-dba2-4255-8ca8-40a8420ec09d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.789183] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1049.789440] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1049.789625] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Deleting the datastore file [datastore2] b412d3e5-0536-413c-9519-5f85c7647cdf {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.790557] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-164386bd-4089-4042-a448-bdd017a0034a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.805272] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The volume has not been displaced from its original location: [datastore1] volume-77787cc2-a082-403d-aca0-bc1b4253df4e/volume-77787cc2-a082-403d-aca0-bc1b4253df4e.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1049.810666] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Reconfiguring VM instance instance-00000059 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1049.812459] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a54c1b0-3ef6-4e18-bda0-fa60c0e77e9e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.826155] env[65522]: DEBUG oslo_vmware.api [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for the task: (returnval){ [ 1049.826155] env[65522]: value = "task-5114654" [ 1049.826155] env[65522]: _type = "Task" [ 1049.826155] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.834214] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "refresh_cache-f1ce2d19-347a-4830-8501-8cceb87d4bf8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.834782] env[65522]: DEBUG nova.compute.manager [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Instance network_info: |[{"id": "76e82556-fba8-4c08-aa9b-1b60d15660c1", "address": "fa:16:3e:07:04:53", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76e82556-fb", "ovs_interfaceid": "76e82556-fba8-4c08-aa9b-1b60d15660c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1049.835393] env[65522]: DEBUG oslo_vmware.api [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1049.835393] env[65522]: value = "task-5114655" [ 1049.835393] env[65522]: _type = "Task" [ 1049.835393] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.835902] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:04:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5e1049e8-c06b-4c93-a9e1-2cbb530f3f95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '76e82556-fba8-4c08-aa9b-1b60d15660c1', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1049.845582] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1049.846815] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1049.854067] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-818728d5-f94c-4f0f-a78c-d8dd16059776 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.869309] env[65522]: DEBUG oslo_vmware.api [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114654, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.871099] env[65522]: DEBUG nova.objects.instance [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lazy-loading 'numa_topology' on Instance uuid 79dd37b5-6b30-48ab-9f00-78214cbd132d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.880348] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.880348] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1049.880348] env[65522]: DEBUG nova.network.neutron [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1049.880925] env[65522]: DEBUG oslo_vmware.api [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114655, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.883224] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1049.883224] env[65522]: value = "task-5114656" [ 1049.883224] env[65522]: _type = "Task" [ 1049.883224] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.892866] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114656, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.976252] env[65522]: DEBUG nova.compute.manager [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1050.199432] env[65522]: DEBUG oslo_vmware.api [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114651, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.240646] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Releasing lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1050.241171] env[65522]: DEBUG nova.compute.manager [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Received event network-changed-d443e903-a2fa-4c64-b185-81fe54af293a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1050.241171] env[65522]: DEBUG nova.compute.manager [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Refreshing instance network info cache due to event network-changed-d443e903-a2fa-4c64-b185-81fe54af293a. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1050.241652] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Acquiring lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.241652] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Acquired lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1050.241948] env[65522]: DEBUG nova.network.neutron [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Refreshing network info cache for port d443e903-a2fa-4c64-b185-81fe54af293a {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1050.278330] env[65522]: DEBUG nova.compute.manager [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Received event network-changed-76e82556-fba8-4c08-aa9b-1b60d15660c1 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1050.278330] env[65522]: DEBUG nova.compute.manager [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Refreshing instance network info cache due to event network-changed-76e82556-fba8-4c08-aa9b-1b60d15660c1. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1050.278330] env[65522]: DEBUG oslo_concurrency.lockutils [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] Acquiring lock "refresh_cache-f1ce2d19-347a-4830-8501-8cceb87d4bf8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.278330] env[65522]: DEBUG oslo_concurrency.lockutils [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] Acquired lock "refresh_cache-f1ce2d19-347a-4830-8501-8cceb87d4bf8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1050.278330] env[65522]: DEBUG nova.network.neutron [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Refreshing network info cache for port 76e82556-fba8-4c08-aa9b-1b60d15660c1 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1050.337515] env[65522]: DEBUG oslo_vmware.api [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Task: {'id': task-5114654, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.111365} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.337851] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.338078] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1050.338304] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1050.338555] env[65522]: INFO nova.compute.manager [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1050.338838] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1050.339097] env[65522]: DEBUG nova.compute.manager [-] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1050.339228] env[65522]: DEBUG nova.network.neutron [-] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1050.339724] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.340478] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1050.340478] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1050.358087] env[65522]: DEBUG oslo_vmware.api [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114655, 'name': ReconfigVM_Task, 'duration_secs': 0.242406} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.358383] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Reconfigured VM instance instance-00000059 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1050.364332] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d04662a9-173f-406f-90c9-4d81960cdd76 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.375138] env[65522]: DEBUG nova.network.neutron [-] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1050.375474] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.377330] env[65522]: INFO nova.compute.claims [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1050.382420] env[65522]: WARNING neutronclient.v2_0.client [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.383114] env[65522]: WARNING openstack [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1050.383471] env[65522]: WARNING openstack [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1050.393578] env[65522]: DEBUG oslo_vmware.api [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1050.393578] env[65522]: value = "task-5114657" [ 1050.393578] env[65522]: _type = "Task" [ 1050.393578] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.408652] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114656, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.413105] env[65522]: DEBUG oslo_vmware.api [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114657, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.505159] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1050.539928] env[65522]: WARNING openstack [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1050.540391] env[65522]: WARNING openstack [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1050.614663] env[65522]: WARNING neutronclient.v2_0.client [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.615499] env[65522]: WARNING openstack [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1050.616062] env[65522]: WARNING openstack [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1050.703960] env[65522]: DEBUG oslo_vmware.api [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114651, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.715209] env[65522]: DEBUG nova.network.neutron [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance_info_cache with network_info: [{"id": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "address": "fa:16:3e:69:05:d7", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf16c93a-75", "ovs_interfaceid": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1050.745032] env[65522]: WARNING neutronclient.v2_0.client [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.745032] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1050.745390] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1050.781359] env[65522]: WARNING neutronclient.v2_0.client [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.782141] env[65522]: WARNING openstack [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1050.782568] env[65522]: WARNING openstack [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1050.868697] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1050.869090] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1050.880589] env[65522]: DEBUG nova.network.neutron [-] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1050.903256] env[65522]: WARNING openstack [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1050.903632] env[65522]: WARNING openstack [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1050.916915] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114656, 'name': CreateVM_Task, 'duration_secs': 0.779316} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.917408] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1050.917875] env[65522]: WARNING neutronclient.v2_0.client [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.918249] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.918403] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1050.918782] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1050.919080] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98a8cd51-f3f9-492b-ae05-f41c25adb40b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.923740] env[65522]: DEBUG oslo_vmware.api [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114657, 'name': ReconfigVM_Task, 'duration_secs': 0.283417} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.924406] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994904', 'volume_id': '77787cc2-a082-403d-aca0-bc1b4253df4e', 'name': 'volume-77787cc2-a082-403d-aca0-bc1b4253df4e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '7a358133-88a5-4dd7-ab69-45b8172870a5', 'attached_at': '', 'detached_at': '', 'volume_id': '77787cc2-a082-403d-aca0-bc1b4253df4e', 'serial': '77787cc2-a082-403d-aca0-bc1b4253df4e'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1050.924694] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1050.925730] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f166be-2eb6-43f3-8266-7e8af821738b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.929335] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1050.929335] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e63e6c-b044-2f1b-e273-4881628d5417" [ 1050.929335] env[65522]: _type = "Task" [ 1050.929335] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.934811] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1050.935402] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9661c48b-2439-4e2e-923a-60347881709f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.938013] env[65522]: WARNING neutronclient.v2_0.client [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.938674] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1050.939300] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1050.950091] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e63e6c-b044-2f1b-e273-4881628d5417, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.990180] env[65522]: WARNING neutronclient.v2_0.client [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.991281] env[65522]: WARNING openstack [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1050.991443] env[65522]: WARNING openstack [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1051.006029] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1051.006307] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1051.006499] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleting the datastore file [datastore2] 7a358133-88a5-4dd7-ab69-45b8172870a5 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1051.007250] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-123a67fa-4b53-4d80-93df-e6620e8d77ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.016273] env[65522]: DEBUG oslo_vmware.api [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1051.016273] env[65522]: value = "task-5114659" [ 1051.016273] env[65522]: _type = "Task" [ 1051.016273] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.026129] env[65522]: DEBUG oslo_vmware.api [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114659, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.044288] env[65522]: DEBUG nova.network.neutron [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updated VIF entry in instance network info cache for port d443e903-a2fa-4c64-b185-81fe54af293a. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1051.044711] env[65522]: DEBUG nova.network.neutron [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updating instance_info_cache with network_info: [{"id": "d443e903-a2fa-4c64-b185-81fe54af293a", "address": "fa:16:3e:81:ed:9d", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd443e903-a2", "ovs_interfaceid": "d443e903-a2fa-4c64-b185-81fe54af293a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1051.088613] env[65522]: DEBUG nova.network.neutron [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Updated VIF entry in instance network info cache for port 76e82556-fba8-4c08-aa9b-1b60d15660c1. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1051.088990] env[65522]: DEBUG nova.network.neutron [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Updating instance_info_cache with network_info: [{"id": "76e82556-fba8-4c08-aa9b-1b60d15660c1", "address": "fa:16:3e:07:04:53", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76e82556-fb", "ovs_interfaceid": "76e82556-fba8-4c08-aa9b-1b60d15660c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1051.196512] env[65522]: DEBUG nova.compute.manager [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1051.197296] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351c8ee0-74cb-4459-93f3-44bc58c01fd7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.204989] env[65522]: DEBUG oslo_vmware.api [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114651, 'name': ReconfigVM_Task, 'duration_secs': 1.043284} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.206707] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Reconfigured VM instance instance-00000045 to attach disk [datastore1] volume-a517fee4-4f9a-40e5-9c5c-0240aebef0cc/volume-a517fee4-4f9a-40e5-9c5c-0240aebef0cc.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.213895] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb971e8d-e8a2-4098-897d-dd5d268e100d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.225629] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1051.232676] env[65522]: DEBUG oslo_vmware.api [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1051.232676] env[65522]: value = "task-5114660" [ 1051.232676] env[65522]: _type = "Task" [ 1051.232676] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.243452] env[65522]: DEBUG oslo_vmware.api [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114660, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.384995] env[65522]: INFO nova.compute.manager [-] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Took 1.05 seconds to deallocate network for instance. [ 1051.444239] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1051.444506] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Processing image e85cd8c4-7ac5-4359-bcda-17d18e3e60f9 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1051.444744] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.444891] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1051.445078] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1051.445341] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-42333904-c791-486e-829f-489779b75789 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.455722] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1051.455927] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1051.456755] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a736d154-c238-413e-81b6-52954007e892 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.464756] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1051.464756] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c282a7-fa83-d14e-fcaa-bc60995db38f" [ 1051.464756] env[65522]: _type = "Task" [ 1051.464756] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.477639] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c282a7-fa83-d14e-fcaa-bc60995db38f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.527612] env[65522]: DEBUG oslo_vmware.api [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114659, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173232} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.527886] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1051.528083] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1051.528267] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1051.528439] env[65522]: INFO nova.compute.manager [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Took 2.34 seconds to destroy the instance on the hypervisor. [ 1051.528692] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1051.528891] env[65522]: DEBUG nova.compute.manager [-] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1051.528992] env[65522]: DEBUG nova.network.neutron [-] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1051.529316] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1051.529875] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1051.530148] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1051.552514] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Releasing lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1051.552786] env[65522]: DEBUG nova.compute.manager [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Received event network-changed-2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1051.552950] env[65522]: DEBUG nova.compute.manager [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Refreshing instance network info cache due to event network-changed-2b2de45d-4c08-4bb4-810a-2f26266cbc20. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1051.553173] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Acquiring lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.553311] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Acquired lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1051.553471] env[65522]: DEBUG nova.network.neutron [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Refreshing network info cache for port 2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1051.572529] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1051.591609] env[65522]: DEBUG oslo_concurrency.lockutils [req-f41a12a1-2beb-415f-89de-d875be75f152 req-b078ae14-4c8f-49d8-899a-5e90950234df service nova] Releasing lock "refresh_cache-f1ce2d19-347a-4830-8501-8cceb87d4bf8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1051.626025] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15f08f0-6a7f-4ad6-be5a-27bbdf36bf0d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.634416] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63a2ccd-37b6-43ec-9845-a790899e6317 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.666282] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3562144c-d62f-4e1d-b8cb-34aa6dcc94ae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.674652] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa1f8b1-f3c0-409c-819a-62783b016233 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.688848] env[65522]: DEBUG nova.compute.provider_tree [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.727613] env[65522]: INFO nova.compute.manager [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] instance snapshotting [ 1051.728481] env[65522]: DEBUG nova.objects.instance [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'flavor' on Instance uuid 9f342b89-bde2-4c35-ae42-cfe1e6973b74 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1051.745097] env[65522]: DEBUG oslo_vmware.api [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114660, 'name': ReconfigVM_Task, 'duration_secs': 0.174785} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.745795] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994906', 'volume_id': 'a517fee4-4f9a-40e5-9c5c-0240aebef0cc', 'name': 'volume-a517fee4-4f9a-40e5-9c5c-0240aebef0cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5ce4a286-efd1-4bbc-a23b-931c6701cfe4', 'attached_at': '', 'detached_at': '', 'volume_id': 'a517fee4-4f9a-40e5-9c5c-0240aebef0cc', 'serial': 'a517fee4-4f9a-40e5-9c5c-0240aebef0cc'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1051.896327] env[65522]: DEBUG oslo_concurrency.lockutils [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1051.977789] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Preparing fetch location {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1051.978165] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Fetch image to [datastore1] OSTACK_IMG_7b3e8aba-cd9f-4ab0-9b9a-ffd1248d7328/OSTACK_IMG_7b3e8aba-cd9f-4ab0-9b9a-ffd1248d7328.vmdk {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1051.978451] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Downloading stream optimized image e85cd8c4-7ac5-4359-bcda-17d18e3e60f9 to [datastore1] OSTACK_IMG_7b3e8aba-cd9f-4ab0-9b9a-ffd1248d7328/OSTACK_IMG_7b3e8aba-cd9f-4ab0-9b9a-ffd1248d7328.vmdk on the data store datastore1 as vApp {{(pid=65522) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1051.978710] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Downloading image file data e85cd8c4-7ac5-4359-bcda-17d18e3e60f9 to the ESX as VM named 'OSTACK_IMG_7b3e8aba-cd9f-4ab0-9b9a-ffd1248d7328' {{(pid=65522) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1052.055850] env[65522]: WARNING neutronclient.v2_0.client [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1052.056876] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1052.056980] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1052.066594] env[65522]: DEBUG oslo_vmware.rw_handles [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1052.066594] env[65522]: value = "resgroup-9" [ 1052.066594] env[65522]: _type = "ResourcePool" [ 1052.066594] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1052.066978] env[65522]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-9040923d-e09e-4512-b9d0-d11ecd58088b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.094113] env[65522]: DEBUG oslo_vmware.rw_handles [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lease: (returnval){ [ 1052.094113] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ed9ac7-adf6-5959-f37d-101471c7d8ee" [ 1052.094113] env[65522]: _type = "HttpNfcLease" [ 1052.094113] env[65522]: } obtained for vApp import into resource pool (val){ [ 1052.094113] env[65522]: value = "resgroup-9" [ 1052.094113] env[65522]: _type = "ResourcePool" [ 1052.094113] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1052.094698] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the lease: (returnval){ [ 1052.094698] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ed9ac7-adf6-5959-f37d-101471c7d8ee" [ 1052.094698] env[65522]: _type = "HttpNfcLease" [ 1052.094698] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1052.105022] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1052.105022] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ed9ac7-adf6-5959-f37d-101471c7d8ee" [ 1052.105022] env[65522]: _type = "HttpNfcLease" [ 1052.105022] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1052.189233] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1052.189341] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1052.197439] env[65522]: DEBUG nova.scheduler.client.report [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1052.244729] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f47a25d-df6d-4d60-883f-66ad96123aad {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.266893] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a264cc2-e393-425a-bf27-25e5fa939f38 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.288272] env[65522]: WARNING neutronclient.v2_0.client [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1052.288975] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1052.289493] env[65522]: WARNING openstack [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1052.305865] env[65522]: DEBUG nova.network.neutron [-] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1052.322641] env[65522]: DEBUG nova.compute.manager [req-3d4ef429-d2b5-42e6-869a-de8aca7f9699 req-81522bdb-0a2c-43a8-9a64-f95bdedf9174 service nova] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Received event network-vif-deleted-0831bbe0-c665-4be1-932f-e4ed9c52eb4e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1052.322641] env[65522]: INFO nova.compute.manager [req-3d4ef429-d2b5-42e6-869a-de8aca7f9699 req-81522bdb-0a2c-43a8-9a64-f95bdedf9174 service nova] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Neutron deleted interface 0831bbe0-c665-4be1-932f-e4ed9c52eb4e; detaching it from the instance and deleting it from the info cache [ 1052.322641] env[65522]: DEBUG nova.network.neutron [req-3d4ef429-d2b5-42e6-869a-de8aca7f9699 req-81522bdb-0a2c-43a8-9a64-f95bdedf9174 service nova] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1052.414849] env[65522]: DEBUG nova.network.neutron [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updated VIF entry in instance network info cache for port 2b2de45d-4c08-4bb4-810a-2f26266cbc20. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1052.415259] env[65522]: DEBUG nova.network.neutron [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updating instance_info_cache with network_info: [{"id": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "address": "fa:16:3e:e9:bd:f6", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b2de45d-4c", "ovs_interfaceid": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1052.606103] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1052.606103] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ed9ac7-adf6-5959-f37d-101471c7d8ee" [ 1052.606103] env[65522]: _type = "HttpNfcLease" [ 1052.606103] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1052.703310] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.348s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.703900] env[65522]: WARNING neutronclient.v2_0.client [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1052.708035] env[65522]: DEBUG oslo_concurrency.lockutils [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 6.892s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1052.708307] env[65522]: DEBUG nova.objects.instance [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1052.745890] env[65522]: INFO nova.network.neutron [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Updating port bd432f78-055d-4cd1-b502-8e058700c03a with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1052.753536] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83c571a-d7ec-436b-b517-6f8413b3032b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.776390] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance '7c6f8218-602d-44f3-8012-de5a96972785' progress to 0 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1052.782201] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1052.783082] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-dbaea650-0a6e-42f2-9289-2a544a094273 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.794009] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1052.794009] env[65522]: value = "task-5114663" [ 1052.794009] env[65522]: _type = "Task" [ 1052.794009] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.809807] env[65522]: DEBUG nova.objects.instance [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lazy-loading 'flavor' on Instance uuid 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.812859] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114663, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.814426] env[65522]: INFO nova.compute.manager [-] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Took 1.28 seconds to deallocate network for instance. [ 1052.825452] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2a170e5-6d33-4bf7-9c17-145b67f8fb71 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.840257] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1346191-d8d4-4a5a-8efe-ca0459becb77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.883837] env[65522]: DEBUG nova.compute.manager [req-3d4ef429-d2b5-42e6-869a-de8aca7f9699 req-81522bdb-0a2c-43a8-9a64-f95bdedf9174 service nova] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Detach interface failed, port_id=0831bbe0-c665-4be1-932f-e4ed9c52eb4e, reason: Instance 7a358133-88a5-4dd7-ab69-45b8172870a5 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1052.919983] env[65522]: DEBUG oslo_concurrency.lockutils [req-5f67ea81-9f2f-4db6-a7d1-fac717d63f66 req-7ddd3ae7-70db-437c-a467-4064308914a1 service nova] Releasing lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1053.107987] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1053.107987] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ed9ac7-adf6-5959-f37d-101471c7d8ee" [ 1053.107987] env[65522]: _type = "HttpNfcLease" [ 1053.107987] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1053.285326] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1053.285823] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e27d28b4-61c8-4b7d-b817-dd15eb4bbe69 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.295862] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1053.295862] env[65522]: value = "task-5114664" [ 1053.295862] env[65522]: _type = "Task" [ 1053.295862] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.309327] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114664, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.313346] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114663, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.318220] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55bf63b1-7c39-4df2-a576-5cad352c41e8 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.896s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1053.359330] env[65522]: DEBUG oslo_concurrency.lockutils [None req-eb59a0a3-78a2-4710-8f94-81e850c23951 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1053.360023] env[65522]: DEBUG oslo_concurrency.lockutils [None req-eb59a0a3-78a2-4710-8f94-81e850c23951 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1053.360023] env[65522]: DEBUG nova.compute.manager [None req-eb59a0a3-78a2-4710-8f94-81e850c23951 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1053.361145] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517d2b66-bf03-462f-8e03-c925046a06d0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.368949] env[65522]: INFO nova.compute.manager [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Took 0.55 seconds to detach 1 volumes for instance. [ 1053.372618] env[65522]: DEBUG nova.compute.manager [None req-eb59a0a3-78a2-4710-8f94-81e850c23951 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65522) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1053.374155] env[65522]: DEBUG nova.objects.instance [None req-eb59a0a3-78a2-4710-8f94-81e850c23951 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lazy-loading 'flavor' on Instance uuid 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.607789] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1053.607789] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ed9ac7-adf6-5959-f37d-101471c7d8ee" [ 1053.607789] env[65522]: _type = "HttpNfcLease" [ 1053.607789] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1053.608277] env[65522]: DEBUG oslo_vmware.rw_handles [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1053.608277] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ed9ac7-adf6-5959-f37d-101471c7d8ee" [ 1053.608277] env[65522]: _type = "HttpNfcLease" [ 1053.608277] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1053.609514] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8715e486-08dd-455c-971a-7892edecbd15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.622501] env[65522]: DEBUG oslo_vmware.rw_handles [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ac752f-6211-c5bd-30ac-80817e3e57a6/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1053.622864] env[65522]: DEBUG oslo_vmware.rw_handles [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ac752f-6211-c5bd-30ac-80817e3e57a6/disk-0.vmdk. {{(pid=65522) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1053.716611] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f3aa3905-7fec-4462-9a09-23131b54c5ac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.720975] env[65522]: DEBUG oslo_concurrency.lockutils [None req-101fb2f0-ee9f-404f-a90c-f3fe56ed0d08 tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1053.722618] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.041s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1053.724973] env[65522]: INFO nova.compute.claims [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1053.810505] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114663, 'name': CreateSnapshot_Task, 'duration_secs': 0.839007} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.811504] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1053.812262] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c3a345b-a2fe-4624-a0d2-b071e798801b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.819142] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114664, 'name': PowerOffVM_Task, 'duration_secs': 0.381239} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.819960] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1053.820334] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance '7c6f8218-602d-44f3-8012-de5a96972785' progress to 17 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1053.880768] env[65522]: DEBUG oslo_concurrency.lockutils [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1054.335763] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1054.336184] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1054.336184] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1054.336413] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1054.336576] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1054.336931] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1054.336931] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1054.337106] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1054.337247] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1054.337411] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1054.337625] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1054.353261] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1054.353698] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6992c4ba-6e44-4cf6-84b3-17e222f69f69 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.370734] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a43ae8ad-dfd2-421f-84a1-8f7d32a20de5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.376029] env[65522]: DEBUG nova.compute.manager [req-34797245-8d6b-47e4-bed5-655da157f7e7 req-5d26a271-ece7-4cdb-839f-12a66b02386f service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Received event network-vif-plugged-bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1054.376295] env[65522]: DEBUG oslo_concurrency.lockutils [req-34797245-8d6b-47e4-bed5-655da157f7e7 req-5d26a271-ece7-4cdb-839f-12a66b02386f service nova] Acquiring lock "79dd37b5-6b30-48ab-9f00-78214cbd132d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1054.376516] env[65522]: DEBUG oslo_concurrency.lockutils [req-34797245-8d6b-47e4-bed5-655da157f7e7 req-5d26a271-ece7-4cdb-839f-12a66b02386f service nova] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.376691] env[65522]: DEBUG oslo_concurrency.lockutils [req-34797245-8d6b-47e4-bed5-655da157f7e7 req-5d26a271-ece7-4cdb-839f-12a66b02386f service nova] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.376859] env[65522]: DEBUG nova.compute.manager [req-34797245-8d6b-47e4-bed5-655da157f7e7 req-5d26a271-ece7-4cdb-839f-12a66b02386f service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] No waiting events found dispatching network-vif-plugged-bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1054.377032] env[65522]: WARNING nova.compute.manager [req-34797245-8d6b-47e4-bed5-655da157f7e7 req-5d26a271-ece7-4cdb-839f-12a66b02386f service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Received unexpected event network-vif-plugged-bd432f78-055d-4cd1-b502-8e058700c03a for instance with vm_state shelved_offloaded and task_state spawning. [ 1054.388160] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb59a0a3-78a2-4710-8f94-81e850c23951 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1054.388825] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1054.388825] env[65522]: value = "task-5114666" [ 1054.388825] env[65522]: _type = "Task" [ 1054.388825] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.392213] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe3b128b-7e40-42f7-9c8f-86a5dad923f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.394736] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1054.394736] env[65522]: value = "task-5114667" [ 1054.394736] env[65522]: _type = "Task" [ 1054.394736] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.413024] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114667, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.418607] env[65522]: DEBUG oslo_vmware.api [None req-eb59a0a3-78a2-4710-8f94-81e850c23951 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1054.418607] env[65522]: value = "task-5114668" [ 1054.418607] env[65522]: _type = "Task" [ 1054.418607] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.419340] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114666, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.431822] env[65522]: DEBUG oslo_vmware.api [None req-eb59a0a3-78a2-4710-8f94-81e850c23951 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114668, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.442664] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.442869] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1054.443067] env[65522]: DEBUG nova.network.neutron [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1054.575485] env[65522]: DEBUG oslo_vmware.rw_handles [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Completed reading data from the image iterator. {{(pid=65522) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1054.575991] env[65522]: DEBUG oslo_vmware.rw_handles [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ac752f-6211-c5bd-30ac-80817e3e57a6/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1054.577025] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bcb5ccd-67b7-47e9-ad74-65eaedd2f716 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.584657] env[65522]: DEBUG oslo_vmware.rw_handles [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ac752f-6211-c5bd-30ac-80817e3e57a6/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1054.584841] env[65522]: DEBUG oslo_vmware.rw_handles [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ac752f-6211-c5bd-30ac-80817e3e57a6/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1054.585486] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-0ba301ab-cf0d-4279-b306-19dd56f343b8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.909960] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114666, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.917944] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114667, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.934025] env[65522]: DEBUG oslo_vmware.api [None req-eb59a0a3-78a2-4710-8f94-81e850c23951 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114668, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.945429] env[65522]: WARNING neutronclient.v2_0.client [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1054.946168] env[65522]: WARNING openstack [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1054.946527] env[65522]: WARNING openstack [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.959454] env[65522]: DEBUG oslo_vmware.rw_handles [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ac752f-6211-c5bd-30ac-80817e3e57a6/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1054.959846] env[65522]: INFO nova.virt.vmwareapi.images [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Downloaded image file data e85cd8c4-7ac5-4359-bcda-17d18e3e60f9 [ 1054.961111] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bd8c698-e3b4-4f94-bca6-7c7625fade7b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.985191] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8dc06a4b-128f-4115-adaf-44b412f908b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.022980] env[65522]: INFO nova.virt.vmwareapi.images [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] The imported VM was unregistered [ 1055.026100] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Caching image {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1055.026399] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Creating directory with path [datastore1] devstack-image-cache_base/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9 {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1055.026738] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22619cfc-615a-4d4b-9d2d-2a5a7af4f354 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.044730] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c5222f-209d-4371-8982-6bfb8e451c2d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.048721] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Created directory with path [datastore1] devstack-image-cache_base/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9 {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1055.048904] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_7b3e8aba-cd9f-4ab0-9b9a-ffd1248d7328/OSTACK_IMG_7b3e8aba-cd9f-4ab0-9b9a-ffd1248d7328.vmdk to [datastore1] devstack-image-cache_base/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9.vmdk. {{(pid=65522) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1055.049598] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-eecfedda-f5e4-45f1-8320-6adf1a3cd35c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.059979] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1beea334-7895-4a7c-bb63-c9c0920d63c7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.064704] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1055.064704] env[65522]: value = "task-5114670" [ 1055.064704] env[65522]: _type = "Task" [ 1055.064704] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.094525] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6565c5-e033-4671-8edd-582ee0f48081 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.102955] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114670, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.106751] env[65522]: WARNING openstack [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1055.107261] env[65522]: WARNING openstack [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1055.118070] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51eb56e8-d2ff-464d-8ad9-64bdd61e5aa2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.136311] env[65522]: DEBUG nova.compute.provider_tree [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.178284] env[65522]: WARNING neutronclient.v2_0.client [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1055.180084] env[65522]: WARNING openstack [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1055.180084] env[65522]: WARNING openstack [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1055.266572] env[65522]: DEBUG nova.network.neutron [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Updating instance_info_cache with network_info: [{"id": "bd432f78-055d-4cd1-b502-8e058700c03a", "address": "fa:16:3e:38:8a:d1", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd432f78-05", "ovs_interfaceid": "bd432f78-055d-4cd1-b502-8e058700c03a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1055.407663] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114666, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.411790] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114667, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.435829] env[65522]: DEBUG oslo_vmware.api [None req-eb59a0a3-78a2-4710-8f94-81e850c23951 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114668, 'name': PowerOffVM_Task, 'duration_secs': 0.55771} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.437442] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb59a0a3-78a2-4710-8f94-81e850c23951 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1055.437730] env[65522]: DEBUG nova.compute.manager [None req-eb59a0a3-78a2-4710-8f94-81e850c23951 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1055.438819] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-408e0e7d-2b25-4111-a048-a05ad9d26c21 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.580153] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114670, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.640351] env[65522]: DEBUG nova.scheduler.client.report [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1055.769025] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1055.805822] env[65522]: DEBUG nova.virt.hardware [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='272ee46793118a9267972025d8dbda22',container_format='bare',created_at=2025-12-12T19:05:05Z,direct_url=,disk_format='vmdk',id=b0c7a88b-7872-412f-9204-61eaa49a2c35,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1297482126-shelved',owner='1546bbdca07c443d84abea1155cfde37',properties=ImageMetaProps,protected=,size=31670784,status='active',tags=,updated_at=2025-12-12T19:05:23Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1055.806108] env[65522]: DEBUG nova.virt.hardware [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1055.806272] env[65522]: DEBUG nova.virt.hardware [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1055.806469] env[65522]: DEBUG nova.virt.hardware [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1055.806617] env[65522]: DEBUG nova.virt.hardware [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1055.806761] env[65522]: DEBUG nova.virt.hardware [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1055.806998] env[65522]: DEBUG nova.virt.hardware [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1055.807199] env[65522]: DEBUG nova.virt.hardware [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1055.807369] env[65522]: DEBUG nova.virt.hardware [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1055.807533] env[65522]: DEBUG nova.virt.hardware [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1055.807704] env[65522]: DEBUG nova.virt.hardware [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1055.808747] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18cf5460-fbe2-4d29-b182-b3fa267ebc6e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.820438] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2cf5f62-e446-47e2-a9b7-9eaa0966855c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.842370] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:8a:d1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4c5eb94-841c-4713-985a-8fc4117fbaf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd432f78-055d-4cd1-b502-8e058700c03a', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1055.850887] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1055.851340] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1055.851608] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a02735ff-620b-483a-ba8d-1c536105b9b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.875127] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1055.875127] env[65522]: value = "task-5114671" [ 1055.875127] env[65522]: _type = "Task" [ 1055.875127] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.886810] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114671, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.908727] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114666, 'name': ReconfigVM_Task, 'duration_secs': 1.25612} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.908727] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance '7c6f8218-602d-44f3-8012-de5a96972785' progress to 33 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1055.916975] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114667, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.955335] env[65522]: DEBUG oslo_concurrency.lockutils [None req-eb59a0a3-78a2-4710-8f94-81e850c23951 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.596s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1056.083033] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114670, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.147307] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1056.148032] env[65522]: DEBUG nova.compute.manager [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1056.151722] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.646s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1056.155022] env[65522]: INFO nova.compute.claims [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1056.389612] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114671, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.392312] env[65522]: DEBUG nova.compute.manager [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Received event network-changed-bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1056.392624] env[65522]: DEBUG nova.compute.manager [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Refreshing instance network info cache due to event network-changed-bd432f78-055d-4cd1-b502-8e058700c03a. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1056.392787] env[65522]: DEBUG oslo_concurrency.lockutils [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] Acquiring lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.392927] env[65522]: DEBUG oslo_concurrency.lockutils [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] Acquired lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1056.393103] env[65522]: DEBUG nova.network.neutron [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Refreshing network info cache for port bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1056.420396] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1056.421807] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1056.421807] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1056.421807] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1056.421807] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1056.421807] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1056.421807] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1056.421807] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1056.422317] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1056.422317] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1056.422317] env[65522]: DEBUG nova.virt.hardware [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1056.428219] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Reconfiguring VM instance instance-0000000a to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1056.429118] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114667, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.429650] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbc61ab8-8dd3-414a-88c4-61518c13c988 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.455507] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1056.455507] env[65522]: value = "task-5114673" [ 1056.455507] env[65522]: _type = "Task" [ 1056.455507] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.470082] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114673, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.580807] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114670, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.657777] env[65522]: DEBUG nova.compute.utils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1056.661993] env[65522]: DEBUG nova.compute.manager [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1056.662253] env[65522]: DEBUG nova.network.neutron [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1056.662659] env[65522]: WARNING neutronclient.v2_0.client [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1056.662953] env[65522]: WARNING neutronclient.v2_0.client [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1056.663649] env[65522]: WARNING openstack [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1056.663956] env[65522]: WARNING openstack [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1056.731606] env[65522]: DEBUG nova.policy [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b7e48c7c3604176b4f844bcd255cb01', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9db9e23ad1b745afa496de2deae83865', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1056.892650] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114671, 'name': CreateVM_Task, 'duration_secs': 0.660501} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.892954] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1056.893682] env[65522]: WARNING neutronclient.v2_0.client [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1056.894498] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b0c7a88b-7872-412f-9204-61eaa49a2c35" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.894799] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b0c7a88b-7872-412f-9204-61eaa49a2c35" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1056.895414] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/b0c7a88b-7872-412f-9204-61eaa49a2c35" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1056.896096] env[65522]: WARNING neutronclient.v2_0.client [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1056.897086] env[65522]: WARNING openstack [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1056.897651] env[65522]: WARNING openstack [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1056.908222] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0899d103-e19f-4529-8a53-d2ddbc3dae60 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.920867] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1056.920867] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521943b2-d7c9-40ef-94cb-af2b8a5a5bf0" [ 1056.920867] env[65522]: _type = "Task" [ 1056.920867] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.925811] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114667, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.938685] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521943b2-d7c9-40ef-94cb-af2b8a5a5bf0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.969786] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114673, 'name': ReconfigVM_Task, 'duration_secs': 0.248618} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.970023] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Reconfigured VM instance instance-0000000a to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1056.971202] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3feca37b-6442-4955-a4fc-e0f65625715f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.000806] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Reconfiguring VM instance instance-0000000a to attach disk [datastore2] 7c6f8218-602d-44f3-8012-de5a96972785/7c6f8218-602d-44f3-8012-de5a96972785.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1057.005729] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16f8f602-f193-4660-9a97-6922fc504f5a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.032165] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1057.032165] env[65522]: value = "task-5114674" [ 1057.032165] env[65522]: _type = "Task" [ 1057.032165] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.047326] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114674, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.077028] env[65522]: WARNING openstack [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.077601] env[65522]: WARNING openstack [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.088049] env[65522]: DEBUG nova.network.neutron [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Successfully created port: 7ba0d5d7-364b-4bea-9e55-b168ec51ad3f {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1057.099714] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114670, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.123525] env[65522]: DEBUG nova.objects.instance [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lazy-loading 'flavor' on Instance uuid 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.163033] env[65522]: DEBUG nova.compute.manager [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1057.167464] env[65522]: WARNING neutronclient.v2_0.client [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1057.168237] env[65522]: WARNING openstack [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.168607] env[65522]: WARNING openstack [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.295347] env[65522]: DEBUG nova.network.neutron [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Updated VIF entry in instance network info cache for port bd432f78-055d-4cd1-b502-8e058700c03a. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1057.295798] env[65522]: DEBUG nova.network.neutron [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Updating instance_info_cache with network_info: [{"id": "bd432f78-055d-4cd1-b502-8e058700c03a", "address": "fa:16:3e:38:8a:d1", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd432f78-05", "ovs_interfaceid": "bd432f78-055d-4cd1-b502-8e058700c03a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1057.427105] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114667, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.440085] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b0c7a88b-7872-412f-9204-61eaa49a2c35" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1057.440413] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Processing image b0c7a88b-7872-412f-9204-61eaa49a2c35 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1057.440674] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/b0c7a88b-7872-412f-9204-61eaa49a2c35/b0c7a88b-7872-412f-9204-61eaa49a2c35.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.441248] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/b0c7a88b-7872-412f-9204-61eaa49a2c35/b0c7a88b-7872-412f-9204-61eaa49a2c35.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1057.441248] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1057.441446] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c643e2be-c0b1-467c-a03e-9eea3de81681 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.446133] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebbfe043-fefc-47c2-85c2-2d0011852fe4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.457197] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eff832be-80fb-485a-b589-c6540f1de3ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.463267] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1057.463510] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1057.464768] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db1b191e-6ac1-4768-a191-2bf90bcc032a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.502192] env[65522]: DEBUG nova.network.neutron [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Successfully created port: ddd341a0-f73c-448c-9ff4-6ef227807e03 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1057.506439] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c5f73c-b8c5-4199-a70c-db2bf2e976da {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.509431] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1057.509431] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52de2712-7f64-893e-d5f0-cd542f78ec2b" [ 1057.509431] env[65522]: _type = "Task" [ 1057.509431] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.523826] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20aa3972-e92e-4952-8e84-cf829b21000e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.531476] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Preparing fetch location {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1057.531476] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Fetch image to [datastore2] OSTACK_IMG_fd557324-2e00-42e9-9e20-64d649c3cfa8/OSTACK_IMG_fd557324-2e00-42e9-9e20-64d649c3cfa8.vmdk {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1057.531476] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Downloading stream optimized image b0c7a88b-7872-412f-9204-61eaa49a2c35 to [datastore2] OSTACK_IMG_fd557324-2e00-42e9-9e20-64d649c3cfa8/OSTACK_IMG_fd557324-2e00-42e9-9e20-64d649c3cfa8.vmdk on the data store datastore2 as vApp {{(pid=65522) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1057.531733] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Downloading image file data b0c7a88b-7872-412f-9204-61eaa49a2c35 to the ESX as VM named 'OSTACK_IMG_fd557324-2e00-42e9-9e20-64d649c3cfa8' {{(pid=65522) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1057.549709] env[65522]: DEBUG nova.compute.provider_tree [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.553423] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114674, 'name': ReconfigVM_Task, 'duration_secs': 0.466041} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.554094] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Reconfigured VM instance instance-0000000a to attach disk [datastore2] 7c6f8218-602d-44f3-8012-de5a96972785/7c6f8218-602d-44f3-8012-de5a96972785.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1057.554809] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance '7c6f8218-602d-44f3-8012-de5a96972785' progress to 50 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1057.601832] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114670, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.630223] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.630223] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquired lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1057.630223] env[65522]: DEBUG nova.network.neutron [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1057.630540] env[65522]: DEBUG nova.objects.instance [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lazy-loading 'info_cache' on Instance uuid 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.641299] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1057.641299] env[65522]: value = "resgroup-9" [ 1057.641299] env[65522]: _type = "ResourcePool" [ 1057.641299] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1057.641706] env[65522]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-9bcd816f-8b31-49e6-908c-cfa3ef9329ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.669525] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lease: (returnval){ [ 1057.669525] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ab45df-8343-8f91-bbec-2d4470a79344" [ 1057.669525] env[65522]: _type = "HttpNfcLease" [ 1057.669525] env[65522]: } obtained for vApp import into resource pool (val){ [ 1057.669525] env[65522]: value = "resgroup-9" [ 1057.669525] env[65522]: _type = "ResourcePool" [ 1057.669525] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1057.669525] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the lease: (returnval){ [ 1057.669525] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ab45df-8343-8f91-bbec-2d4470a79344" [ 1057.669525] env[65522]: _type = "HttpNfcLease" [ 1057.669525] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1057.676768] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1057.676768] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ab45df-8343-8f91-bbec-2d4470a79344" [ 1057.676768] env[65522]: _type = "HttpNfcLease" [ 1057.676768] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1057.798621] env[65522]: DEBUG oslo_concurrency.lockutils [req-d1f4fd4c-8d57-40ea-a1d4-e4f174a95c30 req-8ce9898c-038b-4bf6-9258-90dca29c4553 service nova] Releasing lock "refresh_cache-79dd37b5-6b30-48ab-9f00-78214cbd132d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1057.923803] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114667, 'name': CloneVM_Task, 'duration_secs': 3.516862} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.924086] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Created linked-clone VM from snapshot [ 1057.924845] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70a65f0d-d881-4c01-a888-ec75b5db0814 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.932925] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Uploading image c64abcc1-1c3f-4b83-a6c1-5a8864327b3b {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1057.954601] env[65522]: DEBUG oslo_vmware.rw_handles [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1057.954601] env[65522]: value = "vm-994913" [ 1057.954601] env[65522]: _type = "VirtualMachine" [ 1057.954601] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1057.954921] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ebe36a2d-6106-4a7b-9051-063f8f1c20d8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.962553] env[65522]: DEBUG oslo_vmware.rw_handles [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lease: (returnval){ [ 1057.962553] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5296e86a-761b-7434-b03b-57e85b95e8d8" [ 1057.962553] env[65522]: _type = "HttpNfcLease" [ 1057.962553] env[65522]: } obtained for exporting VM: (result){ [ 1057.962553] env[65522]: value = "vm-994913" [ 1057.962553] env[65522]: _type = "VirtualMachine" [ 1057.962553] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1057.962875] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the lease: (returnval){ [ 1057.962875] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5296e86a-761b-7434-b03b-57e85b95e8d8" [ 1057.962875] env[65522]: _type = "HttpNfcLease" [ 1057.962875] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1057.971450] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1057.971450] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5296e86a-761b-7434-b03b-57e85b95e8d8" [ 1057.971450] env[65522]: _type = "HttpNfcLease" [ 1057.971450] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1058.055478] env[65522]: DEBUG nova.scheduler.client.report [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1058.061276] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cec214d-b086-40b6-b668-963f6b8d548a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.086589] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d4f49e-ccb0-4b14-b413-fa5b8214fea7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.095454] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114670, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.672681} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.109456] env[65522]: INFO nova.virt.vmwareapi.ds_util [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_7b3e8aba-cd9f-4ab0-9b9a-ffd1248d7328/OSTACK_IMG_7b3e8aba-cd9f-4ab0-9b9a-ffd1248d7328.vmdk to [datastore1] devstack-image-cache_base/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9.vmdk. [ 1058.109959] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Cleaning up location [datastore1] OSTACK_IMG_7b3e8aba-cd9f-4ab0-9b9a-ffd1248d7328 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1058.109959] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_7b3e8aba-cd9f-4ab0-9b9a-ffd1248d7328 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1058.110605] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance '7c6f8218-602d-44f3-8012-de5a96972785' progress to 67 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1058.114496] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "interface-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-b6097e79-9e37-4e69-b0f3-40ddbe12d95a" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.114746] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-b6097e79-9e37-4e69-b0f3-40ddbe12d95a" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.115097] env[65522]: DEBUG nova.objects.instance [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'flavor' on Instance uuid 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.116207] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-33f45927-a931-4f2b-b12b-e64311a27678 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.124970] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1058.124970] env[65522]: value = "task-5114678" [ 1058.124970] env[65522]: _type = "Task" [ 1058.124970] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.136524] env[65522]: DEBUG nova.objects.base [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Object Instance<5ce4a286-efd1-4bbc-a23b-931c6701cfe4> lazy-loaded attributes: flavor,info_cache {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1058.138877] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114678, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.177558] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1058.177558] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ab45df-8343-8f91-bbec-2d4470a79344" [ 1058.177558] env[65522]: _type = "HttpNfcLease" [ 1058.177558] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1058.189126] env[65522]: DEBUG nova.compute.manager [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1058.212937] env[65522]: DEBUG nova.virt.hardware [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1058.213364] env[65522]: DEBUG nova.virt.hardware [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1058.213533] env[65522]: DEBUG nova.virt.hardware [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1058.213796] env[65522]: DEBUG nova.virt.hardware [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1058.213945] env[65522]: DEBUG nova.virt.hardware [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1058.214176] env[65522]: DEBUG nova.virt.hardware [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1058.214403] env[65522]: DEBUG nova.virt.hardware [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1058.214564] env[65522]: DEBUG nova.virt.hardware [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1058.214801] env[65522]: DEBUG nova.virt.hardware [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1058.215052] env[65522]: DEBUG nova.virt.hardware [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1058.215236] env[65522]: DEBUG nova.virt.hardware [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1058.217050] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a1f0f2e-0788-4c23-909c-3fe3678a64df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.225595] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e6e523-bf9b-425e-aa27-db3a82e646c2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.471753] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1058.471753] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5296e86a-761b-7434-b03b-57e85b95e8d8" [ 1058.471753] env[65522]: _type = "HttpNfcLease" [ 1058.471753] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1058.472207] env[65522]: DEBUG oslo_vmware.rw_handles [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1058.472207] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5296e86a-761b-7434-b03b-57e85b95e8d8" [ 1058.472207] env[65522]: _type = "HttpNfcLease" [ 1058.472207] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1058.472922] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf40657-79fc-47af-9c29-278fbbd6050c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.483081] env[65522]: DEBUG oslo_vmware.rw_handles [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d89eb7-c071-733d-7fc7-1639d03c9018/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1058.483437] env[65522]: DEBUG oslo_vmware.rw_handles [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d89eb7-c071-733d-7fc7-1639d03c9018/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1058.561339] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1058.561920] env[65522]: DEBUG nova.compute.manager [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1058.565186] env[65522]: DEBUG oslo_concurrency.lockutils [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.669s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.565449] env[65522]: DEBUG nova.objects.instance [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Lazy-loading 'resources' on Instance uuid b412d3e5-0536-413c-9519-5f85c7647cdf {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.620950] env[65522]: WARNING neutronclient.v2_0.client [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.624352] env[65522]: WARNING neutronclient.v2_0.client [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.624967] env[65522]: WARNING openstack [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.625337] env[65522]: WARNING openstack [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.644898] env[65522]: WARNING neutronclient.v2_0.client [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.645644] env[65522]: WARNING openstack [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.646014] env[65522]: WARNING openstack [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.653305] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114678, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.047545} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.654125] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1058.654385] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1058.654592] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9.vmdk to [datastore1] f1ce2d19-347a-4830-8501-8cceb87d4bf8/f1ce2d19-347a-4830-8501-8cceb87d4bf8.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1058.654868] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9585ba79-a6de-47a0-8778-5651d90fa228 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.662415] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1058.662415] env[65522]: value = "task-5114679" [ 1058.662415] env[65522]: _type = "Task" [ 1058.662415] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.671573] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114679, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.678579] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1058.678579] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ab45df-8343-8f91-bbec-2d4470a79344" [ 1058.678579] env[65522]: _type = "HttpNfcLease" [ 1058.678579] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1058.683343] env[65522]: DEBUG nova.network.neutron [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Port cf16c93a-755c-435b-8c86-833d8ea8afbf binding to destination host cpu-1 is already ACTIVE {{(pid=65522) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 1058.820977] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c73b1fd1-78ec-4ca4-8123-28c61413c891 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.922992] env[65522]: DEBUG nova.objects.instance [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'pci_requests' on Instance uuid 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.957056] env[65522]: WARNING openstack [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.957056] env[65522]: WARNING openstack [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.031162] env[65522]: WARNING neutronclient.v2_0.client [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.032071] env[65522]: WARNING openstack [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.032392] env[65522]: WARNING openstack [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.069700] env[65522]: DEBUG nova.compute.utils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1059.073650] env[65522]: DEBUG nova.compute.manager [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1059.073931] env[65522]: DEBUG nova.network.neutron [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1059.074285] env[65522]: WARNING neutronclient.v2_0.client [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.074603] env[65522]: WARNING neutronclient.v2_0.client [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.080098] env[65522]: WARNING openstack [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.080492] env[65522]: WARNING openstack [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.140197] env[65522]: DEBUG nova.policy [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a97b2f232f7844c993dccdc386ce3b2c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9614f18e01594bf083e619c11760fe36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1059.167144] env[65522]: DEBUG nova.network.neutron [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Successfully updated port: 7ba0d5d7-364b-4bea-9e55-b168ec51ad3f {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1059.186027] env[65522]: DEBUG nova.network.neutron [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Updating instance_info_cache with network_info: [{"id": "c2eab4eb-5a45-4c96-a145-bea8b314c82d", "address": "fa:16:3e:22:77:b4", "network": {"id": "f49e9b61-6927-4a5b-a12d-f54f408d42b4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1397158171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a80f979f3dc0477e9462b47f7aa87f14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2eab4eb-5a", "ovs_interfaceid": "c2eab4eb-5a45-4c96-a145-bea8b314c82d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1059.198542] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114679, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.205169] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1059.205169] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ab45df-8343-8f91-bbec-2d4470a79344" [ 1059.205169] env[65522]: _type = "HttpNfcLease" [ 1059.205169] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1059.208839] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1059.208839] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ab45df-8343-8f91-bbec-2d4470a79344" [ 1059.208839] env[65522]: _type = "HttpNfcLease" [ 1059.208839] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1059.210441] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55be4c3-b29f-4120-9ed3-328f4ae46d59 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.226435] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521595bf-70be-c965-4b36-bfe48b042658/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1059.226725] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Creating HTTP connection to write to file with size = 31670784 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521595bf-70be-c965-4b36-bfe48b042658/disk-0.vmdk. {{(pid=65522) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1059.302265] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-92bf0601-3efb-4ddb-bbb3-3f51088b68c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.426530] env[65522]: DEBUG nova.objects.base [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Object Instance<11ef190f-8ade-4705-b3b0-a8ff8b97bcd3> lazy-loaded attributes: flavor,pci_requests {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1059.427544] env[65522]: DEBUG nova.network.neutron [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1059.427544] env[65522]: WARNING neutronclient.v2_0.client [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.427833] env[65522]: WARNING neutronclient.v2_0.client [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.428581] env[65522]: WARNING openstack [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.429489] env[65522]: WARNING openstack [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.453526] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a60be4-efa3-48c1-bd07-84f25cf200a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.460824] env[65522]: DEBUG nova.compute.manager [req-d3ef0aa3-3b66-413c-9a99-ead113b0f449 req-2534e7ed-f8e2-447e-892f-d341849384a4 service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Received event network-vif-plugged-7ba0d5d7-364b-4bea-9e55-b168ec51ad3f {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1059.460824] env[65522]: DEBUG oslo_concurrency.lockutils [req-d3ef0aa3-3b66-413c-9a99-ead113b0f449 req-2534e7ed-f8e2-447e-892f-d341849384a4 service nova] Acquiring lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1059.461489] env[65522]: DEBUG oslo_concurrency.lockutils [req-d3ef0aa3-3b66-413c-9a99-ead113b0f449 req-2534e7ed-f8e2-447e-892f-d341849384a4 service nova] Lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1059.461489] env[65522]: DEBUG oslo_concurrency.lockutils [req-d3ef0aa3-3b66-413c-9a99-ead113b0f449 req-2534e7ed-f8e2-447e-892f-d341849384a4 service nova] Lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1059.461489] env[65522]: DEBUG nova.compute.manager [req-d3ef0aa3-3b66-413c-9a99-ead113b0f449 req-2534e7ed-f8e2-447e-892f-d341849384a4 service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] No waiting events found dispatching network-vif-plugged-7ba0d5d7-364b-4bea-9e55-b168ec51ad3f {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1059.463998] env[65522]: WARNING nova.compute.manager [req-d3ef0aa3-3b66-413c-9a99-ead113b0f449 req-2534e7ed-f8e2-447e-892f-d341849384a4 service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Received unexpected event network-vif-plugged-7ba0d5d7-364b-4bea-9e55-b168ec51ad3f for instance with vm_state building and task_state spawning. [ 1059.472028] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e068aaa-2cbd-4ca6-9426-41a2e30633b4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.525068] env[65522]: DEBUG nova.policy [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b491939798e4481fb433ffb81f366b25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '068c2387de8c406194d9b1762c7292a7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1059.528673] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74467bb1-43a0-4175-8f3f-7056726ad4a1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.542490] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5816b10a-36c0-40fd-998d-06ff632a17fe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.563665] env[65522]: DEBUG nova.compute.provider_tree [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.575237] env[65522]: DEBUG nova.compute.manager [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1059.625680] env[65522]: DEBUG nova.network.neutron [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Successfully created port: 900d53cf-d483-46ab-8a04-c25218fc407c {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1059.683405] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114679, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.693634] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Releasing lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1059.721976] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "7c6f8218-602d-44f3-8012-de5a96972785-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1059.722270] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "7c6f8218-602d-44f3-8012-de5a96972785-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1059.722492] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "7c6f8218-602d-44f3-8012-de5a96972785-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.069786] env[65522]: DEBUG nova.scheduler.client.report [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1060.179832] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114679, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.577514] env[65522]: DEBUG oslo_concurrency.lockutils [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.012s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.583980] env[65522]: DEBUG oslo_concurrency.lockutils [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.703s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1060.584665] env[65522]: DEBUG nova.objects.instance [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lazy-loading 'resources' on Instance uuid 7a358133-88a5-4dd7-ab69-45b8172870a5 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.587575] env[65522]: DEBUG nova.compute.manager [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1060.642017] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Completed reading data from the image iterator. {{(pid=65522) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1060.642017] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521595bf-70be-c965-4b36-bfe48b042658/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1060.643143] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559d72f9-2138-4c0b-89bd-7f9231a925f0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.655874] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521595bf-70be-c965-4b36-bfe48b042658/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1060.656122] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521595bf-70be-c965-4b36-bfe48b042658/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1060.656502] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-faad04e0-b748-4426-bcfe-7b3a5b01da53 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.701569] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1060.701947] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5e07ace-c689-4799-9fa1-8ac8fd601f5c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.728374] env[65522]: WARNING neutronclient.v2_0.client [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1060.834776] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114679, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.836340] env[65522]: DEBUG oslo_vmware.api [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1060.836340] env[65522]: value = "task-5114681" [ 1060.836340] env[65522]: _type = "Task" [ 1060.836340] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.848751] env[65522]: INFO nova.scheduler.client.report [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Deleted allocations for instance b412d3e5-0536-413c-9519-5f85c7647cdf [ 1060.855782] env[65522]: DEBUG oslo_vmware.api [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114681, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.859270] env[65522]: DEBUG nova.virt.hardware [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1060.859646] env[65522]: DEBUG nova.virt.hardware [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1060.860085] env[65522]: DEBUG nova.virt.hardware [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1060.860445] env[65522]: DEBUG nova.virt.hardware [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1060.860762] env[65522]: DEBUG nova.virt.hardware [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1060.861447] env[65522]: DEBUG nova.virt.hardware [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1060.861762] env[65522]: DEBUG nova.virt.hardware [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1060.862108] env[65522]: DEBUG nova.virt.hardware [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1060.862379] env[65522]: DEBUG nova.virt.hardware [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1060.862678] env[65522]: DEBUG nova.virt.hardware [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1060.863051] env[65522]: DEBUG nova.virt.hardware [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1060.868303] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1ccbcc-722e-4cc3-9818-11a5f49fed39 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.881488] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26e9568-0c69-450b-8d14-c746abe9c778 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.890085] env[65522]: DEBUG oslo_vmware.rw_handles [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/521595bf-70be-c965-4b36-bfe48b042658/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1060.890830] env[65522]: INFO nova.virt.vmwareapi.images [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Downloaded image file data b0c7a88b-7872-412f-9204-61eaa49a2c35 [ 1060.891509] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa208db-e605-4aed-880f-c4219d431b44 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.925587] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23fc0570-790d-4974-ad3a-26fa98c1adec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.954921] env[65522]: INFO nova.virt.vmwareapi.images [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] The imported VM was unregistered [ 1060.958204] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Caching image {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1060.958493] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Creating directory with path [datastore2] devstack-image-cache_base/b0c7a88b-7872-412f-9204-61eaa49a2c35 {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1060.958837] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c14f555b-db2b-425b-a630-a6ada5690814 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.998020] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Created directory with path [datastore2] devstack-image-cache_base/b0c7a88b-7872-412f-9204-61eaa49a2c35 {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1060.998020] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_fd557324-2e00-42e9-9e20-64d649c3cfa8/OSTACK_IMG_fd557324-2e00-42e9-9e20-64d649c3cfa8.vmdk to [datastore2] devstack-image-cache_base/b0c7a88b-7872-412f-9204-61eaa49a2c35/b0c7a88b-7872-412f-9204-61eaa49a2c35.vmdk. {{(pid=65522) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1060.998020] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-3415d283-a803-4f4d-964f-50489ef03a41 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.009721] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1061.009721] env[65522]: value = "task-5114683" [ 1061.009721] env[65522]: _type = "Task" [ 1061.009721] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.024012] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114683, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.173026] env[65522]: DEBUG nova.network.neutron [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Successfully updated port: b6097e79-9e37-4e69-b0f3-40ddbe12d95a {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1061.182629] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114679, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.245108] env[65522]: DEBUG nova.network.neutron [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Successfully updated port: 900d53cf-d483-46ab-8a04-c25218fc407c {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1061.290381] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.290526] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1061.290673] env[65522]: DEBUG nova.network.neutron [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1061.356416] env[65522]: DEBUG oslo_vmware.api [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114681, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.375832] env[65522]: DEBUG oslo_concurrency.lockutils [None req-89b4c054-b15b-481e-9717-9f71481eb06e tempest-ServersListShow296Test-183371836 tempest-ServersListShow296Test-183371836-project-member] Lock "b412d3e5-0536-413c-9519-5f85c7647cdf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.761s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1061.403486] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af0ac90c-005c-431c-8792-44fb4c24edb1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.416924] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022458ec-887e-497b-878d-3e5fba58918f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.460697] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beef07d8-3869-4cef-9016-ba7c1a3912e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.473414] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac63abf-dabf-4315-a6c3-216ded78e607 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.491781] env[65522]: DEBUG nova.compute.provider_tree [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.526030] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114683, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.677247] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114679, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.961957} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.677962] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9/e85cd8c4-7ac5-4359-bcda-17d18e3e60f9.vmdk to [datastore1] f1ce2d19-347a-4830-8501-8cceb87d4bf8/f1ce2d19-347a-4830-8501-8cceb87d4bf8.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1061.679216] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0808117d-99f7-42f9-888a-775387d67763 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.685044] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.685231] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1061.685784] env[65522]: DEBUG nova.network.neutron [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1061.709053] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] f1ce2d19-347a-4830-8501-8cceb87d4bf8/f1ce2d19-347a-4830-8501-8cceb87d4bf8.vmdk or device None with type streamOptimized {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1061.711231] env[65522]: WARNING openstack [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.711794] env[65522]: WARNING openstack [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.719133] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7ec0e48-9e4a-4ac3-9c54-1b11f63ae956 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.745491] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1061.745491] env[65522]: value = "task-5114684" [ 1061.745491] env[65522]: _type = "Task" [ 1061.745491] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.750026] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "refresh_cache-210184da-48db-4926-ae00-7c0f29622810" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.750317] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "refresh_cache-210184da-48db-4926-ae00-7c0f29622810" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1061.750565] env[65522]: DEBUG nova.network.neutron [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1061.761047] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114684, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.794464] env[65522]: WARNING neutronclient.v2_0.client [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1061.795360] env[65522]: WARNING openstack [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.795785] env[65522]: WARNING openstack [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.806168] env[65522]: WARNING nova.network.neutron [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] 1ea0ec37-ea40-4911-b6ad-54375253cc42 already exists in list: networks containing: ['1ea0ec37-ea40-4911-b6ad-54375253cc42']. ignoring it [ 1061.858955] env[65522]: DEBUG oslo_vmware.api [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114681, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.910211] env[65522]: WARNING openstack [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1061.911068] env[65522]: WARNING openstack [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1061.995373] env[65522]: DEBUG nova.scheduler.client.report [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1062.025998] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114683, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.073392] env[65522]: WARNING openstack [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.074238] env[65522]: WARNING openstack [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.128159] env[65522]: WARNING neutronclient.v2_0.client [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.129137] env[65522]: WARNING openstack [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.129617] env[65522]: WARNING openstack [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.168540] env[65522]: DEBUG nova.compute.manager [req-54c8f2f1-e67e-4b95-8521-4d8dee6e704f req-f137b432-2208-40fa-9c73-513a239e2076 service nova] [instance: 210184da-48db-4926-ae00-7c0f29622810] Received event network-vif-plugged-900d53cf-d483-46ab-8a04-c25218fc407c {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1062.169970] env[65522]: DEBUG oslo_concurrency.lockutils [req-54c8f2f1-e67e-4b95-8521-4d8dee6e704f req-f137b432-2208-40fa-9c73-513a239e2076 service nova] Acquiring lock "210184da-48db-4926-ae00-7c0f29622810-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1062.169970] env[65522]: DEBUG oslo_concurrency.lockutils [req-54c8f2f1-e67e-4b95-8521-4d8dee6e704f req-f137b432-2208-40fa-9c73-513a239e2076 service nova] Lock "210184da-48db-4926-ae00-7c0f29622810-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1062.169970] env[65522]: DEBUG oslo_concurrency.lockutils [req-54c8f2f1-e67e-4b95-8521-4d8dee6e704f req-f137b432-2208-40fa-9c73-513a239e2076 service nova] Lock "210184da-48db-4926-ae00-7c0f29622810-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1062.169970] env[65522]: DEBUG nova.compute.manager [req-54c8f2f1-e67e-4b95-8521-4d8dee6e704f req-f137b432-2208-40fa-9c73-513a239e2076 service nova] [instance: 210184da-48db-4926-ae00-7c0f29622810] No waiting events found dispatching network-vif-plugged-900d53cf-d483-46ab-8a04-c25218fc407c {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1062.169970] env[65522]: WARNING nova.compute.manager [req-54c8f2f1-e67e-4b95-8521-4d8dee6e704f req-f137b432-2208-40fa-9c73-513a239e2076 service nova] [instance: 210184da-48db-4926-ae00-7c0f29622810] Received unexpected event network-vif-plugged-900d53cf-d483-46ab-8a04-c25218fc407c for instance with vm_state building and task_state spawning. [ 1062.208446] env[65522]: DEBUG nova.compute.manager [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Received event network-changed-7ba0d5d7-364b-4bea-9e55-b168ec51ad3f {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1062.208446] env[65522]: DEBUG nova.compute.manager [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Refreshing instance network info cache due to event network-changed-7ba0d5d7-364b-4bea-9e55-b168ec51ad3f. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1062.208446] env[65522]: DEBUG oslo_concurrency.lockutils [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Acquiring lock "refresh_cache-c071e816-fe74-4aae-9f8e-7a1c69431f9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.208446] env[65522]: DEBUG oslo_concurrency.lockutils [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Acquired lock "refresh_cache-c071e816-fe74-4aae-9f8e-7a1c69431f9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1062.208446] env[65522]: DEBUG nova.network.neutron [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Refreshing network info cache for port 7ba0d5d7-364b-4bea-9e55-b168ec51ad3f {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1062.228234] env[65522]: WARNING neutronclient.v2_0.client [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.229203] env[65522]: WARNING openstack [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.229588] env[65522]: WARNING openstack [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.256855] env[65522]: WARNING openstack [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.257613] env[65522]: WARNING openstack [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.272466] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.359926] env[65522]: DEBUG oslo_vmware.api [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114681, 'name': PowerOnVM_Task, 'duration_secs': 1.21594} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.361498] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1062.361700] env[65522]: DEBUG nova.compute.manager [None req-3eb46fd9-6174-4e6f-bb77-62957f751dfe tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1062.362766] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bbea75f-5c66-47d5-83cd-a7bc13f261c4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.382530] env[65522]: DEBUG nova.network.neutron [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1062.458790] env[65522]: WARNING openstack [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.459384] env[65522]: WARNING openstack [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.500471] env[65522]: DEBUG oslo_concurrency.lockutils [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.917s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1062.524551] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114683, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.529542] env[65522]: INFO nova.scheduler.client.report [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleted allocations for instance 7a358133-88a5-4dd7-ab69-45b8172870a5 [ 1062.712496] env[65522]: WARNING neutronclient.v2_0.client [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.715700] env[65522]: WARNING openstack [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.716465] env[65522]: WARNING openstack [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.727807] env[65522]: DEBUG nova.network.neutron [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance_info_cache with network_info: [{"id": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "address": "fa:16:3e:69:05:d7", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf16c93a-75", "ovs_interfaceid": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1062.739823] env[65522]: WARNING openstack [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.740317] env[65522]: WARNING openstack [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.766821] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.893650] env[65522]: DEBUG nova.network.neutron [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Successfully updated port: ddd341a0-f73c-448c-9ff4-6ef227807e03 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1062.937676] env[65522]: DEBUG nova.network.neutron [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1062.963024] env[65522]: WARNING neutronclient.v2_0.client [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.965672] env[65522]: WARNING openstack [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.965672] env[65522]: WARNING openstack [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.993410] env[65522]: WARNING neutronclient.v2_0.client [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.999073] env[65522]: WARNING openstack [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.999770] env[65522]: WARNING openstack [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1063.026790] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114683, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.050297] env[65522]: DEBUG oslo_concurrency.lockutils [None req-37907da5-6c30-490a-8bc5-e165c98b22b5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "7a358133-88a5-4dd7-ab69-45b8172870a5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.376s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1063.099491] env[65522]: DEBUG nova.network.neutron [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updating instance_info_cache with network_info: [{"id": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "address": "fa:16:3e:e9:bd:f6", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b2de45d-4c", "ovs_interfaceid": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b6097e79-9e37-4e69-b0f3-40ddbe12d95a", "address": "fa:16:3e:1b:7b:c6", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6097e79-9e", "ovs_interfaceid": "b6097e79-9e37-4e69-b0f3-40ddbe12d95a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1063.137514] env[65522]: DEBUG nova.network.neutron [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1063.162292] env[65522]: DEBUG nova.network.neutron [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Updating instance_info_cache with network_info: [{"id": "900d53cf-d483-46ab-8a04-c25218fc407c", "address": "fa:16:3e:af:37:68", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap900d53cf-d4", "ovs_interfaceid": "900d53cf-d483-46ab-8a04-c25218fc407c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1063.231535] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1063.264057] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.400035] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "refresh_cache-c071e816-fe74-4aae-9f8e-7a1c69431f9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.524395] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114683, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.603780] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1063.604668] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.605334] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1063.606142] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8abe54-28f4-4c3c-993e-0e5d81179c28 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.629323] env[65522]: DEBUG nova.virt.hardware [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1063.629739] env[65522]: DEBUG nova.virt.hardware [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1063.629988] env[65522]: DEBUG nova.virt.hardware [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1063.630286] env[65522]: DEBUG nova.virt.hardware [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1063.630483] env[65522]: DEBUG nova.virt.hardware [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1063.630643] env[65522]: DEBUG nova.virt.hardware [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1063.631609] env[65522]: DEBUG nova.virt.hardware [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1063.631609] env[65522]: DEBUG nova.virt.hardware [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1063.631609] env[65522]: DEBUG nova.virt.hardware [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1063.632054] env[65522]: DEBUG nova.virt.hardware [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1063.632054] env[65522]: DEBUG nova.virt.hardware [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1063.640027] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Reconfiguring VM to attach interface {{(pid=65522) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1063.640811] env[65522]: DEBUG oslo_concurrency.lockutils [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Releasing lock "refresh_cache-c071e816-fe74-4aae-9f8e-7a1c69431f9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1063.641137] env[65522]: DEBUG nova.compute.manager [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Received event network-vif-plugged-b6097e79-9e37-4e69-b0f3-40ddbe12d95a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1063.641341] env[65522]: DEBUG oslo_concurrency.lockutils [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Acquiring lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1063.641665] env[65522]: DEBUG oslo_concurrency.lockutils [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1063.641846] env[65522]: DEBUG oslo_concurrency.lockutils [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1063.642046] env[65522]: DEBUG nova.compute.manager [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] No waiting events found dispatching network-vif-plugged-b6097e79-9e37-4e69-b0f3-40ddbe12d95a {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1063.642256] env[65522]: WARNING nova.compute.manager [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Received unexpected event network-vif-plugged-b6097e79-9e37-4e69-b0f3-40ddbe12d95a for instance with vm_state active and task_state None. [ 1063.642566] env[65522]: DEBUG nova.compute.manager [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Received event network-changed-b6097e79-9e37-4e69-b0f3-40ddbe12d95a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1063.642683] env[65522]: DEBUG nova.compute.manager [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Refreshing instance network info cache due to event network-changed-b6097e79-9e37-4e69-b0f3-40ddbe12d95a. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1063.642941] env[65522]: DEBUG oslo_concurrency.lockutils [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Acquiring lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.643112] env[65522]: DEBUG oslo_concurrency.lockutils [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Acquired lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1063.643269] env[65522]: DEBUG nova.network.neutron [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Refreshing network info cache for port b6097e79-9e37-4e69-b0f3-40ddbe12d95a {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1063.644644] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9aec753a-4451-43ad-b723-800c1e9ba446 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.658752] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquired lock "refresh_cache-c071e816-fe74-4aae-9f8e-7a1c69431f9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1063.658999] env[65522]: DEBUG nova.network.neutron [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1063.666105] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "refresh_cache-210184da-48db-4926-ae00-7c0f29622810" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1063.667125] env[65522]: DEBUG nova.compute.manager [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Instance network_info: |[{"id": "900d53cf-d483-46ab-8a04-c25218fc407c", "address": "fa:16:3e:af:37:68", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap900d53cf-d4", "ovs_interfaceid": "900d53cf-d483-46ab-8a04-c25218fc407c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1063.670120] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:af:37:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10b81051-1eb1-406b-888c-4548c470c77e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '900d53cf-d483-46ab-8a04-c25218fc407c', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1063.686178] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1063.690120] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 210184da-48db-4926-ae00-7c0f29622810] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1063.690920] env[65522]: DEBUG oslo_vmware.api [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1063.690920] env[65522]: value = "task-5114686" [ 1063.690920] env[65522]: _type = "Task" [ 1063.690920] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.691401] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23b259f5-3b68-48fd-bdfb-c71d93350aa5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.744901] env[65522]: DEBUG oslo_vmware.api [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114686, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.747084] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1063.747084] env[65522]: value = "task-5114687" [ 1063.747084] env[65522]: _type = "Task" [ 1063.747084] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.770500] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114687, 'name': CreateVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.775793] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3a3506-3656-4463-aecf-3792f67bc8ca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.778838] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114684, 'name': ReconfigVM_Task, 'duration_secs': 1.897427} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.779223] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Reconfigured VM instance instance-0000005c to attach disk [datastore1] f1ce2d19-347a-4830-8501-8cceb87d4bf8/f1ce2d19-347a-4830-8501-8cceb87d4bf8.vmdk or device None with type streamOptimized {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1063.780705] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd421bc9-9e49-45c0-8de2-e4edaa3539a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.812963] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b784d8-fa8f-4037-b546-44f2844c5f8d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.816867] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1063.816867] env[65522]: value = "task-5114688" [ 1063.816867] env[65522]: _type = "Task" [ 1063.816867] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.825618] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance '7c6f8218-602d-44f3-8012-de5a96972785' progress to 83 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1063.837644] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114688, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.029070] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114683, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.159501] env[65522]: WARNING neutronclient.v2_0.client [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1064.160146] env[65522]: WARNING openstack [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.161539] env[65522]: WARNING openstack [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.169668] env[65522]: WARNING openstack [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.169874] env[65522]: WARNING openstack [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.237956] env[65522]: DEBUG oslo_vmware.api [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114686, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.264470] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114687, 'name': CreateVM_Task, 'duration_secs': 0.491207} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.264470] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 210184da-48db-4926-ae00-7c0f29622810] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1064.264470] env[65522]: WARNING neutronclient.v2_0.client [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1064.264470] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.264782] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.265491] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1064.265491] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9184ccec-dcc0-4733-9a01-c30ac7850857 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.272287] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1064.272287] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cf2e66-b2f9-01ac-7ba9-5f706d8750e2" [ 1064.272287] env[65522]: _type = "Task" [ 1064.272287] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.284344] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cf2e66-b2f9-01ac-7ba9-5f706d8750e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.301956] env[65522]: DEBUG nova.network.neutron [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1064.338525] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1064.339693] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114688, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.339693] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a56ce96f-2cd7-404f-9e5a-6b15a6d47e7f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.348690] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1064.348690] env[65522]: value = "task-5114689" [ 1064.348690] env[65522]: _type = "Task" [ 1064.348690] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.364318] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114689, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.379335] env[65522]: WARNING openstack [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.380094] env[65522]: WARNING openstack [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.389509] env[65522]: WARNING openstack [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.390518] env[65522]: WARNING openstack [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.487229] env[65522]: WARNING neutronclient.v2_0.client [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1064.487963] env[65522]: WARNING openstack [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.488369] env[65522]: WARNING openstack [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.498125] env[65522]: WARNING neutronclient.v2_0.client [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1064.498790] env[65522]: WARNING openstack [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.499174] env[65522]: WARNING openstack [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.524936] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114683, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.320474} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.525215] env[65522]: INFO nova.virt.vmwareapi.ds_util [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_fd557324-2e00-42e9-9e20-64d649c3cfa8/OSTACK_IMG_fd557324-2e00-42e9-9e20-64d649c3cfa8.vmdk to [datastore2] devstack-image-cache_base/b0c7a88b-7872-412f-9204-61eaa49a2c35/b0c7a88b-7872-412f-9204-61eaa49a2c35.vmdk. [ 1064.525426] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Cleaning up location [datastore2] OSTACK_IMG_fd557324-2e00-42e9-9e20-64d649c3cfa8 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1064.525598] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_fd557324-2e00-42e9-9e20-64d649c3cfa8 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1064.525901] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8ed36d4f-4287-4aa1-95e0-61d33606a25a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.539669] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1064.539669] env[65522]: value = "task-5114690" [ 1064.539669] env[65522]: _type = "Task" [ 1064.539669] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.556682] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114690, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.640194] env[65522]: DEBUG nova.network.neutron [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updated VIF entry in instance network info cache for port b6097e79-9e37-4e69-b0f3-40ddbe12d95a. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1064.640194] env[65522]: DEBUG nova.network.neutron [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updating instance_info_cache with network_info: [{"id": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "address": "fa:16:3e:e9:bd:f6", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b2de45d-4c", "ovs_interfaceid": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b6097e79-9e37-4e69-b0f3-40ddbe12d95a", "address": "fa:16:3e:1b:7b:c6", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6097e79-9e", "ovs_interfaceid": "b6097e79-9e37-4e69-b0f3-40ddbe12d95a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1064.653783] env[65522]: WARNING openstack [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.654399] env[65522]: WARNING openstack [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.667079] env[65522]: DEBUG nova.compute.manager [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] [instance: 210184da-48db-4926-ae00-7c0f29622810] Received event network-changed-900d53cf-d483-46ab-8a04-c25218fc407c {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1064.667386] env[65522]: DEBUG nova.compute.manager [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] [instance: 210184da-48db-4926-ae00-7c0f29622810] Refreshing instance network info cache due to event network-changed-900d53cf-d483-46ab-8a04-c25218fc407c. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1064.667590] env[65522]: DEBUG oslo_concurrency.lockutils [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] Acquiring lock "refresh_cache-210184da-48db-4926-ae00-7c0f29622810" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.668241] env[65522]: DEBUG oslo_concurrency.lockutils [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] Acquired lock "refresh_cache-210184da-48db-4926-ae00-7c0f29622810" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.668241] env[65522]: DEBUG nova.network.neutron [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] [instance: 210184da-48db-4926-ae00-7c0f29622810] Refreshing network info cache for port 900d53cf-d483-46ab-8a04-c25218fc407c {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1064.730049] env[65522]: WARNING neutronclient.v2_0.client [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1064.731098] env[65522]: WARNING openstack [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.732164] env[65522]: WARNING openstack [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.749858] env[65522]: DEBUG oslo_vmware.api [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114686, 'name': ReconfigVM_Task, 'duration_secs': 0.873532} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.750359] env[65522]: WARNING neutronclient.v2_0.client [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1064.750763] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1064.750902] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Reconfigured VM to attach interface {{(pid=65522) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1064.785233] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cf2e66-b2f9-01ac-7ba9-5f706d8750e2, 'name': SearchDatastore_Task, 'duration_secs': 0.021057} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.786676] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1064.786937] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1064.787188] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.787330] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.787507] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1064.795036] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7d924b20-987e-401e-9903-c3c01a4e1099 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.797426] env[65522]: DEBUG oslo_concurrency.lockutils [None req-440b71af-b924-4f26-8ed9-28a61e781076 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "0c1012e5-635c-40d3-bf21-b59ce98ae026" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.797639] env[65522]: DEBUG oslo_concurrency.lockutils [None req-440b71af-b924-4f26-8ed9-28a61e781076 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "0c1012e5-635c-40d3-bf21-b59ce98ae026" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1064.809761] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1064.810050] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1064.811439] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84190fc2-d8ae-4912-a885-b496d97e8102 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.819070] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1064.819070] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d0b01f-3b39-b8c4-7bc7-f0d44ecab452" [ 1064.819070] env[65522]: _type = "Task" [ 1064.819070] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.825454] env[65522]: DEBUG nova.compute.manager [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Received event network-vif-plugged-ddd341a0-f73c-448c-9ff4-6ef227807e03 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1064.825674] env[65522]: DEBUG oslo_concurrency.lockutils [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] Acquiring lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.825900] env[65522]: DEBUG oslo_concurrency.lockutils [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] Lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1064.826042] env[65522]: DEBUG oslo_concurrency.lockutils [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] Lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1064.826206] env[65522]: DEBUG nova.compute.manager [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] No waiting events found dispatching network-vif-plugged-ddd341a0-f73c-448c-9ff4-6ef227807e03 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1064.826413] env[65522]: WARNING nova.compute.manager [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Received unexpected event network-vif-plugged-ddd341a0-f73c-448c-9ff4-6ef227807e03 for instance with vm_state building and task_state spawning. [ 1064.826529] env[65522]: DEBUG nova.compute.manager [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Received event network-changed-ddd341a0-f73c-448c-9ff4-6ef227807e03 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1064.826665] env[65522]: DEBUG nova.compute.manager [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Refreshing instance network info cache due to event network-changed-ddd341a0-f73c-448c-9ff4-6ef227807e03. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1064.826827] env[65522]: DEBUG oslo_concurrency.lockutils [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] Acquiring lock "refresh_cache-c071e816-fe74-4aae-9f8e-7a1c69431f9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.842107] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d0b01f-3b39-b8c4-7bc7-f0d44ecab452, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.842409] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114688, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.844866] env[65522]: DEBUG nova.network.neutron [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Updating instance_info_cache with network_info: [{"id": "7ba0d5d7-364b-4bea-9e55-b168ec51ad3f", "address": "fa:16:3e:cd:96:62", "network": {"id": "b8206df8-18d0-443d-93f1-df8d9c0ae385", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2012759704", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.70", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ba0d5d7-36", "ovs_interfaceid": "7ba0d5d7-364b-4bea-9e55-b168ec51ad3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ddd341a0-f73c-448c-9ff4-6ef227807e03", "address": "fa:16:3e:02:17:ec", "network": {"id": "a7f44549-33cf-4e57-8048-421dbc1e638a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1189055942", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd341a0-f7", "ovs_interfaceid": "ddd341a0-f73c-448c-9ff4-6ef227807e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1064.860326] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114689, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.049654] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114690, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.067525} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.049934] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1065.050114] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/b0c7a88b-7872-412f-9204-61eaa49a2c35/b0c7a88b-7872-412f-9204-61eaa49a2c35.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1065.050370] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/b0c7a88b-7872-412f-9204-61eaa49a2c35/b0c7a88b-7872-412f-9204-61eaa49a2c35.vmdk to [datastore2] 79dd37b5-6b30-48ab-9f00-78214cbd132d/79dd37b5-6b30-48ab-9f00-78214cbd132d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1065.050653] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-568e8718-2692-426c-a6f0-36a300b224e2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.058334] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1065.058334] env[65522]: value = "task-5114691" [ 1065.058334] env[65522]: _type = "Task" [ 1065.058334] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.067395] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114691, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.141594] env[65522]: DEBUG oslo_concurrency.lockutils [req-bde2f2a5-be08-4de3-b84b-8a2f9e80b9a1 req-bdcac227-0e2b-46b4-985e-fa9b0d9ad73d service nova] Releasing lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1065.172742] env[65522]: WARNING neutronclient.v2_0.client [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1065.173536] env[65522]: WARNING openstack [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.174051] env[65522]: WARNING openstack [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.259246] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c3442b46-4606-46d9-9d90-52362f74109b tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-b6097e79-9e37-4e69-b0f3-40ddbe12d95a" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.144s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1065.300364] env[65522]: DEBUG nova.compute.manager [None req-440b71af-b924-4f26-8ed9-28a61e781076 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 0c1012e5-635c-40d3-bf21-b59ce98ae026] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1065.339144] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114688, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.339600] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d0b01f-3b39-b8c4-7bc7-f0d44ecab452, 'name': SearchDatastore_Task, 'duration_secs': 0.040893} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.340960] env[65522]: WARNING openstack [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.341341] env[65522]: WARNING openstack [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.349818] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Releasing lock "refresh_cache-c071e816-fe74-4aae-9f8e-7a1c69431f9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1065.350248] env[65522]: DEBUG nova.compute.manager [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Instance network_info: |[{"id": "7ba0d5d7-364b-4bea-9e55-b168ec51ad3f", "address": "fa:16:3e:cd:96:62", "network": {"id": "b8206df8-18d0-443d-93f1-df8d9c0ae385", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2012759704", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.70", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ba0d5d7-36", "ovs_interfaceid": "7ba0d5d7-364b-4bea-9e55-b168ec51ad3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ddd341a0-f73c-448c-9ff4-6ef227807e03", "address": "fa:16:3e:02:17:ec", "network": {"id": "a7f44549-33cf-4e57-8048-421dbc1e638a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1189055942", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd341a0-f7", "ovs_interfaceid": "ddd341a0-f73c-448c-9ff4-6ef227807e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1065.350935] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e201cab3-60ce-41b4-978d-64eecf86a17a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.353801] env[65522]: DEBUG oslo_concurrency.lockutils [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] Acquired lock "refresh_cache-c071e816-fe74-4aae-9f8e-7a1c69431f9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1065.353968] env[65522]: DEBUG nova.network.neutron [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Refreshing network info cache for port ddd341a0-f73c-448c-9ff4-6ef227807e03 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1065.355484] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cd:96:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd891a662-7da0-4be5-8d0d-01b4ff70552a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7ba0d5d7-364b-4bea-9e55-b168ec51ad3f', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:17:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc6d5964-1106-4345-a26d-185dabd4ff0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ddd341a0-f73c-448c-9ff4-6ef227807e03', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1065.365210] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1065.367193] env[65522]: WARNING neutronclient.v2_0.client [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1065.367859] env[65522]: WARNING openstack [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.368881] env[65522]: WARNING openstack [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.379603] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1065.381516] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89bb5b56-681c-4c10-8a38-464d3967788d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.406831] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1065.406831] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b499b9-593a-2bd3-3f3e-851728cc3c9c" [ 1065.406831] env[65522]: _type = "Task" [ 1065.406831] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.417501] env[65522]: DEBUG oslo_vmware.api [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114689, 'name': PowerOnVM_Task, 'duration_secs': 0.60645} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.418583] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.418783] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7998a260-a28e-4bb8-b7c7-b9677ec58366 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance '7c6f8218-602d-44f3-8012-de5a96972785' progress to 100 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1065.423917] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1065.423917] env[65522]: value = "task-5114692" [ 1065.423917] env[65522]: _type = "Task" [ 1065.423917] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.431506] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b499b9-593a-2bd3-3f3e-851728cc3c9c, 'name': SearchDatastore_Task, 'duration_secs': 0.025751} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.431757] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1065.434078] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 210184da-48db-4926-ae00-7c0f29622810/210184da-48db-4926-ae00-7c0f29622810.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1065.434078] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2cdc240-9627-47ab-9e40-88eab5658526 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.438026] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114692, 'name': CreateVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.448785] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1065.448785] env[65522]: value = "task-5114693" [ 1065.448785] env[65522]: _type = "Task" [ 1065.448785] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.458518] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114693, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.459903] env[65522]: WARNING neutronclient.v2_0.client [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1065.460749] env[65522]: WARNING openstack [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.461196] env[65522]: WARNING openstack [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.518298] env[65522]: WARNING openstack [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.518737] env[65522]: WARNING openstack [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.553334] env[65522]: DEBUG nova.network.neutron [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] [instance: 210184da-48db-4926-ae00-7c0f29622810] Updated VIF entry in instance network info cache for port 900d53cf-d483-46ab-8a04-c25218fc407c. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1065.553712] env[65522]: DEBUG nova.network.neutron [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] [instance: 210184da-48db-4926-ae00-7c0f29622810] Updating instance_info_cache with network_info: [{"id": "900d53cf-d483-46ab-8a04-c25218fc407c", "address": "fa:16:3e:af:37:68", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap900d53cf-d4", "ovs_interfaceid": "900d53cf-d483-46ab-8a04-c25218fc407c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1065.572336] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114691, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.592305] env[65522]: WARNING neutronclient.v2_0.client [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1065.592916] env[65522]: WARNING openstack [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1065.593312] env[65522]: WARNING openstack [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1065.678546] env[65522]: DEBUG nova.network.neutron [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Updated VIF entry in instance network info cache for port ddd341a0-f73c-448c-9ff4-6ef227807e03. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1065.679087] env[65522]: DEBUG nova.network.neutron [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Updating instance_info_cache with network_info: [{"id": "7ba0d5d7-364b-4bea-9e55-b168ec51ad3f", "address": "fa:16:3e:cd:96:62", "network": {"id": "b8206df8-18d0-443d-93f1-df8d9c0ae385", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2012759704", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.70", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d891a662-7da0-4be5-8d0d-01b4ff70552a", "external-id": "nsx-vlan-transportzone-815", "segmentation_id": 815, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7ba0d5d7-36", "ovs_interfaceid": "7ba0d5d7-364b-4bea-9e55-b168ec51ad3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ddd341a0-f73c-448c-9ff4-6ef227807e03", "address": "fa:16:3e:02:17:ec", "network": {"id": "a7f44549-33cf-4e57-8048-421dbc1e638a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1189055942", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "9db9e23ad1b745afa496de2deae83865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd341a0-f7", "ovs_interfaceid": "ddd341a0-f73c-448c-9ff4-6ef227807e03", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1065.809172] env[65522]: DEBUG nova.compute.manager [None req-440b71af-b924-4f26-8ed9-28a61e781076 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 0c1012e5-635c-40d3-bf21-b59ce98ae026] Instance disappeared before build. {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2482}} [ 1065.830862] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114688, 'name': Rename_Task, 'duration_secs': 1.600525} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.831235] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1065.831546] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-03aa5edb-8269-48c9-b296-e4b63cd38faa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.840583] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1065.840583] env[65522]: value = "task-5114694" [ 1065.840583] env[65522]: _type = "Task" [ 1065.840583] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.851884] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114694, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.941835] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114692, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.962256] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114693, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.056934] env[65522]: DEBUG oslo_concurrency.lockutils [req-fbcf57ba-4730-4a25-bc2d-be3701b12e50 req-3a7b7928-49be-410a-95a6-4f1158755b9f service nova] Releasing lock "refresh_cache-210184da-48db-4926-ae00-7c0f29622810" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1066.074652] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114691, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.182567] env[65522]: DEBUG oslo_concurrency.lockutils [req-bb8b7706-13fb-4cae-b2d6-1e62ae55e27d req-3a5d052a-082d-4dd8-9276-08920c0354aa service nova] Releasing lock "refresh_cache-c071e816-fe74-4aae-9f8e-7a1c69431f9d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1066.325148] env[65522]: DEBUG oslo_concurrency.lockutils [None req-440b71af-b924-4f26-8ed9-28a61e781076 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "0c1012e5-635c-40d3-bf21-b59ce98ae026" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 1.527s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1066.353176] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114694, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.439242] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114692, 'name': CreateVM_Task, 'duration_secs': 0.737337} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.439242] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1066.439666] env[65522]: WARNING neutronclient.v2_0.client [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1066.440120] env[65522]: WARNING neutronclient.v2_0.client [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1066.440356] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.440542] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1066.440865] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1066.441190] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ad14c83-677e-4cc9-a8da-e26aca4e47aa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.448372] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1066.448372] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522a105a-704f-d7b6-2345-2b36a01768a6" [ 1066.448372] env[65522]: _type = "Task" [ 1066.448372] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.461368] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114693, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.465147] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522a105a-704f-d7b6-2345-2b36a01768a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.560835] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Acquiring lock "f6980679-bc2e-442f-b424-8cf80de86c02" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1066.561129] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Lock "f6980679-bc2e-442f-b424-8cf80de86c02" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1066.574789] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114691, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.856883] env[65522]: DEBUG oslo_vmware.api [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114694, 'name': PowerOnVM_Task, 'duration_secs': 0.934142} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.857306] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1066.857569] env[65522]: INFO nova.compute.manager [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Took 19.91 seconds to spawn the instance on the hypervisor. [ 1066.857677] env[65522]: DEBUG nova.compute.manager [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1066.858623] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3033d18d-5779-41d7-bce2-00faa8bdd42c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.967141] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114693, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.972754] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522a105a-704f-d7b6-2345-2b36a01768a6, 'name': SearchDatastore_Task, 'duration_secs': 0.10906} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.973624] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1066.973945] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1066.974278] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.974615] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1066.974708] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1066.975827] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-18dff34e-995c-48e5-bd3f-96a4a4ce114a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.999979] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1067.000210] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1067.001023] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b88ec1ed-ab41-4bdc-9054-9a408b9e96ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.009878] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1067.009878] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5232b8cf-961a-bd2c-a358-531ceeb81bee" [ 1067.009878] env[65522]: _type = "Task" [ 1067.009878] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.021944] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5232b8cf-961a-bd2c-a358-531ceeb81bee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.063633] env[65522]: DEBUG nova.compute.manager [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1067.079495] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114691, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.342915] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "interface-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-b6097e79-9e37-4e69-b0f3-40ddbe12d95a" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1067.342915] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-b6097e79-9e37-4e69-b0f3-40ddbe12d95a" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.380754] env[65522]: INFO nova.compute.manager [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Took 34.90 seconds to build instance. [ 1067.470470] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114693, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.482038] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "821a5847-c255-4846-9041-2a0144f4539b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1067.482291] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "821a5847-c255-4846-9041-2a0144f4539b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.521971] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5232b8cf-961a-bd2c-a358-531ceeb81bee, 'name': SearchDatastore_Task, 'duration_secs': 0.097424} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.523203] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-637fffaa-97af-49f1-91e0-29facdcfb43f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.532184] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1067.532184] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c564d6-fee7-5430-6cb5-983814784cbb" [ 1067.532184] env[65522]: _type = "Task" [ 1067.532184] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.542808] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c564d6-fee7-5430-6cb5-983814784cbb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.583425] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114691, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.593249] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1067.593249] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.593911] env[65522]: INFO nova.compute.claims [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1067.846673] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.846914] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1067.848312] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa21f2cb-0316-49be-9c4f-fb85f2cd14b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.872059] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71da508e-8e4c-4184-8f9d-27098d41d3c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.896193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-afbdf363-f30b-40bb-a1e9-1ad138e5900b tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "f1ce2d19-347a-4830-8501-8cceb87d4bf8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.428s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.897195] env[65522]: WARNING neutronclient.v2_0.client [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.903056] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Reconfiguring VM to detach interface {{(pid=65522) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1067.904174] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be41bdb8-757d-433d-9f8c-f1b45a69ace6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.927432] env[65522]: DEBUG oslo_vmware.api [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1067.927432] env[65522]: value = "task-5114695" [ 1067.927432] env[65522]: _type = "Task" [ 1067.927432] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.939463] env[65522]: DEBUG oslo_vmware.api [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114695, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.969406] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114693, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.985198] env[65522]: DEBUG nova.compute.manager [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1068.050343] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c564d6-fee7-5430-6cb5-983814784cbb, 'name': SearchDatastore_Task, 'duration_secs': 0.137127} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.050860] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1068.051287] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] c071e816-fe74-4aae-9f8e-7a1c69431f9d/c071e816-fe74-4aae-9f8e-7a1c69431f9d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1068.051726] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-670d8047-a146-4ff4-b894-51e40bef70bf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.063312] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1068.063312] env[65522]: value = "task-5114696" [ 1068.063312] env[65522]: _type = "Task" [ 1068.063312] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.077297] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114691, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.080768] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114696, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.282221] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "f1ce2d19-347a-4830-8501-8cceb87d4bf8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1068.282681] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "f1ce2d19-347a-4830-8501-8cceb87d4bf8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1068.282923] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "f1ce2d19-347a-4830-8501-8cceb87d4bf8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1068.283146] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "f1ce2d19-347a-4830-8501-8cceb87d4bf8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1068.283323] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "f1ce2d19-347a-4830-8501-8cceb87d4bf8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1068.285912] env[65522]: INFO nova.compute.manager [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Terminating instance [ 1068.317743] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "7c6f8218-602d-44f3-8012-de5a96972785" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1068.317956] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "7c6f8218-602d-44f3-8012-de5a96972785" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1068.318186] env[65522]: DEBUG nova.compute.manager [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Going to confirm migration 5 {{(pid=65522) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1068.445431] env[65522]: DEBUG oslo_vmware.api [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.469218] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114693, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.585364} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.469511] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 210184da-48db-4926-ae00-7c0f29622810/210184da-48db-4926-ae00-7c0f29622810.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1068.469735] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1068.470029] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-24f0cb4e-017b-4424-9f97-17eaaab5a146 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.481065] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1068.481065] env[65522]: value = "task-5114697" [ 1068.481065] env[65522]: _type = "Task" [ 1068.481065] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.495543] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114697, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.513498] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1068.583779] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114691, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.588619] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114696, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.789657] env[65522]: DEBUG nova.compute.manager [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1068.793019] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1068.793019] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f199b86-0bde-4b6b-92a9-73ff4be9b39c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.802303] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1068.802634] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9cac35e3-da0f-4096-8c01-4a727df20d13 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.814020] env[65522]: DEBUG oslo_vmware.api [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1068.814020] env[65522]: value = "task-5114698" [ 1068.814020] env[65522]: _type = "Task" [ 1068.814020] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.829158] env[65522]: DEBUG oslo_vmware.api [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114698, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.829158] env[65522]: WARNING neutronclient.v2_0.client [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1068.851962] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01df323-18f1-4a48-a587-74ce46c45f61 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.860717] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454114e4-1fb5-4ddb-bfe4-57d9f53d9d47 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.865151] env[65522]: WARNING neutronclient.v2_0.client [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1068.865470] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.865621] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1068.865789] env[65522]: DEBUG nova.network.neutron [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1068.865975] env[65522]: DEBUG nova.objects.instance [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lazy-loading 'info_cache' on Instance uuid 7c6f8218-602d-44f3-8012-de5a96972785 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.898083] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5730712-b05c-427a-80b2-91ba4460769c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.909488] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16a2df6-39eb-4c8f-aad3-bd0f201e36d6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.930588] env[65522]: DEBUG nova.compute.provider_tree [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.942423] env[65522]: DEBUG oslo_vmware.api [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.993149] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114697, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091398} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.993527] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1068.995340] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9259dbf-78ed-45e3-8bd6-b01fa63797f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.022973] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 210184da-48db-4926-ae00-7c0f29622810/210184da-48db-4926-ae00-7c0f29622810.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1069.023383] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b687ce67-a07f-4ab2-9406-9d1ed67218ff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.045059] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1069.045059] env[65522]: value = "task-5114699" [ 1069.045059] env[65522]: _type = "Task" [ 1069.045059] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.054435] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114699, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.079038] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114696, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.085825] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114691, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.854828} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.086277] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/b0c7a88b-7872-412f-9204-61eaa49a2c35/b0c7a88b-7872-412f-9204-61eaa49a2c35.vmdk to [datastore2] 79dd37b5-6b30-48ab-9f00-78214cbd132d/79dd37b5-6b30-48ab-9f00-78214cbd132d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1069.087311] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de95f6b-0d45-44fc-bad8-72deff52d798 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.114353] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 79dd37b5-6b30-48ab-9f00-78214cbd132d/79dd37b5-6b30-48ab-9f00-78214cbd132d.vmdk or device None with type streamOptimized {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1069.114766] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3235169b-10a2-49f2-a8ec-d48d459102c2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.136469] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1069.136469] env[65522]: value = "task-5114700" [ 1069.136469] env[65522]: _type = "Task" [ 1069.136469] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.145879] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114700, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.324561] env[65522]: DEBUG oslo_vmware.api [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114698, 'name': PowerOffVM_Task, 'duration_secs': 0.472967} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.324828] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1069.325049] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1069.325446] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c6f29ce0-4eee-42e4-8300-48d34bc71202 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.437316] env[65522]: DEBUG nova.scheduler.client.report [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1069.451909] env[65522]: DEBUG oslo_vmware.api [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.487138] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1069.487468] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1069.487694] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleting the datastore file [datastore1] f1ce2d19-347a-4830-8501-8cceb87d4bf8 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1069.488223] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f17b9474-e2a0-42fa-bfd6-746154b265a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.502359] env[65522]: DEBUG oslo_vmware.api [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1069.502359] env[65522]: value = "task-5114702" [ 1069.502359] env[65522]: _type = "Task" [ 1069.502359] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.514905] env[65522]: DEBUG oslo_vmware.api [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114702, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.556645] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114699, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.578936] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114696, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.649574] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114700, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.873122] env[65522]: WARNING neutronclient.v2_0.client [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1069.873912] env[65522]: WARNING openstack [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1069.874365] env[65522]: WARNING openstack [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1069.941159] env[65522]: DEBUG oslo_vmware.api [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.944108] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1069.944479] env[65522]: DEBUG nova.compute.manager [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1069.947760] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.434s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1069.949253] env[65522]: INFO nova.compute.claims [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1070.003192] env[65522]: WARNING openstack [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1070.003655] env[65522]: WARNING openstack [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1070.023537] env[65522]: DEBUG oslo_vmware.api [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114702, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.056872] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114699, 'name': ReconfigVM_Task, 'duration_secs': 0.72569} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.057161] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 210184da-48db-4926-ae00-7c0f29622810/210184da-48db-4926-ae00-7c0f29622810.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1070.057991] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42d63631-7d81-48a8-b392-09d984431066 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.067056] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1070.067056] env[65522]: value = "task-5114703" [ 1070.067056] env[65522]: _type = "Task" [ 1070.067056] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.076852] env[65522]: WARNING neutronclient.v2_0.client [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1070.077574] env[65522]: WARNING openstack [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1070.077925] env[65522]: WARNING openstack [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1070.093847] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114696, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.610717} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.094049] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114703, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.094366] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] c071e816-fe74-4aae-9f8e-7a1c69431f9d/c071e816-fe74-4aae-9f8e-7a1c69431f9d.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1070.094650] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1070.094933] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c71028dc-5c47-40bc-bbea-01cc8b6f503c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.103890] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1070.103890] env[65522]: value = "task-5114704" [ 1070.103890] env[65522]: _type = "Task" [ 1070.103890] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.115048] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114704, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.148875] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114700, 'name': ReconfigVM_Task, 'duration_secs': 0.671778} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.153927] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 79dd37b5-6b30-48ab-9f00-78214cbd132d/79dd37b5-6b30-48ab-9f00-78214cbd132d.vmdk or device None with type streamOptimized {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1070.155421] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_options': None, 'encrypted': False, 'boot_index': 0, 'disk_bus': None, 'size': 0, 'encryption_format': None, 'guest_format': None, 'encryption_secret_uuid': None, 'device_name': '/dev/sda', 'device_type': 'disk', 'image_id': 'a889db67-7337-4e32-8e34-642f34402926'}], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': False, 'boot_index': None, 'mount_device': '/dev/sdb', 'disk_bus': None, 'guest_format': None, 'device_type': None, 'attachment_id': '980c2a76-ba44-415d-823b-0f303e04d510', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994901', 'volume_id': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'name': 'volume-fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '79dd37b5-6b30-48ab-9f00-78214cbd132d', 'attached_at': '', 'detached_at': '', 'volume_id': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'serial': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc'}, 'volume_type': None}], 'swap': None} {{(pid=65522) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1070.155654] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1070.155867] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994901', 'volume_id': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'name': 'volume-fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '79dd37b5-6b30-48ab-9f00-78214cbd132d', 'attached_at': '', 'detached_at': '', 'volume_id': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'serial': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1070.157365] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7332d47c-488f-4fbf-9682-20de83c28034 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.178169] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5896138-5973-44d7-9273-9554d01fb120 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.208043] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] volume-fd5bf7dd-1d57-4e86-badc-405c9e3774cc/volume-fd5bf7dd-1d57-4e86-badc-405c9e3774cc.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1070.210846] env[65522]: DEBUG nova.network.neutron [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance_info_cache with network_info: [{"id": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "address": "fa:16:3e:69:05:d7", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf16c93a-75", "ovs_interfaceid": "cf16c93a-755c-435b-8c86-833d8ea8afbf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1070.212241] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62f07e30-0e72-4d0c-b052-67b32169544d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.227127] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-7c6f8218-602d-44f3-8012-de5a96972785" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1070.227356] env[65522]: DEBUG nova.objects.instance [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lazy-loading 'migration_context' on Instance uuid 7c6f8218-602d-44f3-8012-de5a96972785 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.233791] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1070.233791] env[65522]: value = "task-5114705" [ 1070.233791] env[65522]: _type = "Task" [ 1070.233791] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.244597] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114705, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.442275] env[65522]: DEBUG oslo_vmware.api [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.454218] env[65522]: DEBUG nova.compute.utils [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1070.455684] env[65522]: DEBUG nova.compute.manager [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1070.455882] env[65522]: DEBUG nova.network.neutron [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1070.457086] env[65522]: WARNING neutronclient.v2_0.client [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1070.457086] env[65522]: WARNING neutronclient.v2_0.client [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1070.457219] env[65522]: WARNING openstack [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1070.457532] env[65522]: WARNING openstack [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1070.521021] env[65522]: DEBUG nova.policy [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd137f4123c494c9c89e2c0fa991c50b9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c4c0ca0038354ab48c5b2cdb004d549e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1070.527456] env[65522]: DEBUG oslo_vmware.api [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114702, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.713172} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.527890] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1070.528073] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1070.528307] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1070.528534] env[65522]: INFO nova.compute.manager [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Took 1.74 seconds to destroy the instance on the hypervisor. [ 1070.528874] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1070.529182] env[65522]: DEBUG nova.compute.manager [-] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1070.529504] env[65522]: DEBUG nova.network.neutron [-] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1070.529663] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1070.530349] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1070.530734] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1070.571256] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1070.583269] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114703, 'name': Rename_Task, 'duration_secs': 0.271776} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.583574] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1070.584503] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4fadbaf5-fd5e-4dcf-861d-ae94ede0c085 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.592505] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1070.592505] env[65522]: value = "task-5114706" [ 1070.592505] env[65522]: _type = "Task" [ 1070.592505] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.601516] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114706, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.613680] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114704, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100071} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.614021] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1070.614868] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d806de6-d188-406c-bbf2-b510bdce3e70 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.641138] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] c071e816-fe74-4aae-9f8e-7a1c69431f9d/c071e816-fe74-4aae-9f8e-7a1c69431f9d.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1070.641526] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61ad045d-3071-459d-9316-3834b63b9816 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.664806] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1070.664806] env[65522]: value = "task-5114707" [ 1070.664806] env[65522]: _type = "Task" [ 1070.664806] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.676325] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114707, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.731096] env[65522]: DEBUG nova.objects.base [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Object Instance<7c6f8218-602d-44f3-8012-de5a96972785> lazy-loaded attributes: info_cache,migration_context {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1070.732565] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc703b7e-582a-43bf-9fe8-9a4152772828 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.768532] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee841025-7795-433e-9c0c-c19b945804d5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.772389] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.776118] env[65522]: DEBUG oslo_vmware.api [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1070.776118] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ac2416-5531-5e7a-b243-b0495719dcec" [ 1070.776118] env[65522]: _type = "Task" [ 1070.776118] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.789312] env[65522]: DEBUG oslo_vmware.api [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ac2416-5531-5e7a-b243-b0495719dcec, 'name': SearchDatastore_Task, 'duration_secs': 0.009981} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.789899] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1070.835101] env[65522]: DEBUG nova.compute.manager [req-ce0488a8-eef3-4cf5-9180-e256d9c4a653 req-a511759b-671d-413c-bc56-2cf80507f46b service nova] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Received event network-vif-deleted-76e82556-fba8-4c08-aa9b-1b60d15660c1 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1070.835101] env[65522]: INFO nova.compute.manager [req-ce0488a8-eef3-4cf5-9180-e256d9c4a653 req-a511759b-671d-413c-bc56-2cf80507f46b service nova] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Neutron deleted interface 76e82556-fba8-4c08-aa9b-1b60d15660c1; detaching it from the instance and deleting it from the info cache [ 1070.835101] env[65522]: DEBUG nova.network.neutron [req-ce0488a8-eef3-4cf5-9180-e256d9c4a653 req-a511759b-671d-413c-bc56-2cf80507f46b service nova] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1070.944034] env[65522]: DEBUG oslo_vmware.api [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.945277] env[65522]: DEBUG nova.network.neutron [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Successfully created port: 01011dd2-f801-46a0-8528-8305d2c08ee9 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1070.970849] env[65522]: DEBUG nova.compute.manager [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1071.108760] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114706, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.179473] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114707, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.247760] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114705, 'name': ReconfigVM_Task, 'duration_secs': 0.67557} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.247959] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Reconfigured VM instance instance-0000004c to attach disk [datastore1] volume-fd5bf7dd-1d57-4e86-badc-405c9e3774cc/volume-fd5bf7dd-1d57-4e86-badc-405c9e3774cc.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.256578] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-abd3f1b3-fd9a-47df-8391-0ab751708640 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.274529] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1071.274529] env[65522]: value = "task-5114708" [ 1071.274529] env[65522]: _type = "Task" [ 1071.274529] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.282348] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c59a0f92-36ab-42f5-a6d3-8ab50b8f0b13 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.290575] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114708, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.297439] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e4bbff-6baf-4c62-b99c-25440deb37e9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.331924] env[65522]: DEBUG nova.network.neutron [-] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1071.335232] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee06fb81-447a-4e65-8aa9-6e17701ebcd1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.346325] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d225724-adac-4338-b3ce-842b562ea3c9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.351832] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e00b3fdd-db99-4ada-8ab8-a5d1dd29cd96 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.370609] env[65522]: DEBUG nova.compute.provider_tree [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.376684] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4fea10-d0c3-4690-b7a1-85a4412cf762 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.424249] env[65522]: DEBUG nova.compute.manager [req-ce0488a8-eef3-4cf5-9180-e256d9c4a653 req-a511759b-671d-413c-bc56-2cf80507f46b service nova] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Detach interface failed, port_id=76e82556-fba8-4c08-aa9b-1b60d15660c1, reason: Instance f1ce2d19-347a-4830-8501-8cceb87d4bf8 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1071.444902] env[65522]: DEBUG oslo_vmware.api [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.477103] env[65522]: INFO nova.virt.block_device [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Booting with volume b4d96581-08cb-4fb8-98a2-bc1957d04206 at /dev/sda [ 1071.522463] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f9d6005-c409-48d4-bee0-6765fed80342 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.535438] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fbf84cb-b279-4c75-9b04-b69ea0e3afbe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.586787] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-878f8bca-a229-4638-8e47-9f35f7ce617a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.602776] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9da920-7025-4f14-ba6d-a07bf06970dc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.620210] env[65522]: DEBUG oslo_vmware.api [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114706, 'name': PowerOnVM_Task, 'duration_secs': 0.854936} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.620516] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1071.620736] env[65522]: INFO nova.compute.manager [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Took 11.03 seconds to spawn the instance on the hypervisor. [ 1071.620926] env[65522]: DEBUG nova.compute.manager [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1071.621922] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852bf3c8-49f2-47c7-a9d1-b1a43fe15632 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.642900] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a28360-a23b-4b9c-a515-0b7f19c3bb2a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.653705] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-576f473a-2fad-46d3-921a-3c1594c9e693 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.673371] env[65522]: DEBUG nova.virt.block_device [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Updating existing volume attachment record: a4e2eeb5-7784-44fc-8393-7219b63b229d {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1071.684219] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114707, 'name': ReconfigVM_Task, 'duration_secs': 0.57058} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.684560] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Reconfigured VM instance instance-0000005d to attach disk [datastore2] c071e816-fe74-4aae-9f8e-7a1c69431f9d/c071e816-fe74-4aae-9f8e-7a1c69431f9d.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.685272] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5dd7ab68-9d57-4a05-967b-50e75c507c40 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.693465] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1071.693465] env[65522]: value = "task-5114709" [ 1071.693465] env[65522]: _type = "Task" [ 1071.693465] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.704088] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114709, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.785840] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114708, 'name': ReconfigVM_Task, 'duration_secs': 0.286269} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.786230] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994901', 'volume_id': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'name': 'volume-fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '79dd37b5-6b30-48ab-9f00-78214cbd132d', 'attached_at': '', 'detached_at': '', 'volume_id': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'serial': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1071.787010] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-188c06b9-5e08-4b0e-ad8f-5cca2dc4c7a3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.794446] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1071.794446] env[65522]: value = "task-5114710" [ 1071.794446] env[65522]: _type = "Task" [ 1071.794446] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.804553] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114710, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.834986] env[65522]: INFO nova.compute.manager [-] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Took 1.31 seconds to deallocate network for instance. [ 1071.876059] env[65522]: DEBUG nova.scheduler.client.report [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1071.945694] env[65522]: DEBUG oslo_vmware.api [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.160734] env[65522]: INFO nova.compute.manager [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Took 21.68 seconds to build instance. [ 1072.204537] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114709, 'name': Rename_Task, 'duration_secs': 0.208826} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.204827] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1072.206237] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-38c49ef7-12f3-4a91-b7e7-270db91f504b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.213047] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1072.213047] env[65522]: value = "task-5114711" [ 1072.213047] env[65522]: _type = "Task" [ 1072.213047] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.224663] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114711, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.306693] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114710, 'name': Rename_Task, 'duration_secs': 0.301857} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.306991] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1072.307271] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9899cb42-f300-4eb4-9f37-8015ab7bf1a9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.315149] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1072.315149] env[65522]: value = "task-5114712" [ 1072.315149] env[65522]: _type = "Task" [ 1072.315149] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.328618] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114712, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.345029] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1072.383266] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1072.383266] env[65522]: DEBUG nova.compute.manager [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1072.385583] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.596s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1072.445765] env[65522]: DEBUG oslo_vmware.api [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.563397] env[65522]: DEBUG nova.network.neutron [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Successfully updated port: 01011dd2-f801-46a0-8528-8305d2c08ee9 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1072.662089] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cab63abf-0b46-409f-bc3d-65c4b41a45cd tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "210184da-48db-4926-ae00-7c0f29622810" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.189s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1072.712231] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "99353848-2f0f-4388-9fcd-91e799342386" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1072.712533] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "99353848-2f0f-4388-9fcd-91e799342386" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1072.730375] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114711, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.829510] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114712, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.889918] env[65522]: DEBUG nova.compute.utils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1072.895662] env[65522]: DEBUG nova.compute.manager [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1072.895748] env[65522]: DEBUG nova.network.neutron [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1072.896474] env[65522]: WARNING neutronclient.v2_0.client [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.896685] env[65522]: WARNING neutronclient.v2_0.client [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.897352] env[65522]: WARNING openstack [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.897727] env[65522]: WARNING openstack [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1072.910915] env[65522]: DEBUG nova.compute.manager [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Received event network-vif-plugged-01011dd2-f801-46a0-8528-8305d2c08ee9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1072.911321] env[65522]: DEBUG oslo_concurrency.lockutils [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] Acquiring lock "f6980679-bc2e-442f-b424-8cf80de86c02-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1072.912204] env[65522]: DEBUG oslo_concurrency.lockutils [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] Lock "f6980679-bc2e-442f-b424-8cf80de86c02-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1072.912204] env[65522]: DEBUG oslo_concurrency.lockutils [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] Lock "f6980679-bc2e-442f-b424-8cf80de86c02-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1072.912204] env[65522]: DEBUG nova.compute.manager [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] No waiting events found dispatching network-vif-plugged-01011dd2-f801-46a0-8528-8305d2c08ee9 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1072.912361] env[65522]: WARNING nova.compute.manager [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Received unexpected event network-vif-plugged-01011dd2-f801-46a0-8528-8305d2c08ee9 for instance with vm_state building and task_state block_device_mapping. [ 1072.912456] env[65522]: DEBUG nova.compute.manager [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Received event network-changed-01011dd2-f801-46a0-8528-8305d2c08ee9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1072.912689] env[65522]: DEBUG nova.compute.manager [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Refreshing instance network info cache due to event network-changed-01011dd2-f801-46a0-8528-8305d2c08ee9. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1072.912911] env[65522]: DEBUG oslo_concurrency.lockutils [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] Acquiring lock "refresh_cache-f6980679-bc2e-442f-b424-8cf80de86c02" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.913122] env[65522]: DEBUG oslo_concurrency.lockutils [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] Acquired lock "refresh_cache-f6980679-bc2e-442f-b424-8cf80de86c02" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1072.913328] env[65522]: DEBUG nova.network.neutron [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Refreshing network info cache for port 01011dd2-f801-46a0-8528-8305d2c08ee9 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1072.948907] env[65522]: DEBUG oslo_vmware.api [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.966922] env[65522]: DEBUG nova.policy [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8f8fa586d8c74bb9805a1148fa7ed037', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f93c790ff61543bd8e134bcf9cb20bb2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1073.066517] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Acquiring lock "refresh_cache-f6980679-bc2e-442f-b424-8cf80de86c02" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.157070] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-641f6cb4-ea09-4ca5-a9b2-5725189da723 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.165584] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b536f8-781d-4b34-b23a-921008fdc714 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.199640] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee8fa9d-7e77-4a7c-a429-21c29304a25c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.208349] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d1a85dd-4bd1-41a5-bba3-387680ff1c5f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.227025] env[65522]: DEBUG nova.compute.utils [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1073.228793] env[65522]: DEBUG nova.compute.provider_tree [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.236995] env[65522]: DEBUG oslo_vmware.api [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114711, 'name': PowerOnVM_Task, 'duration_secs': 0.601158} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.236995] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1073.237177] env[65522]: INFO nova.compute.manager [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Took 15.05 seconds to spawn the instance on the hypervisor. [ 1073.237549] env[65522]: DEBUG nova.compute.manager [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1073.238269] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d58cb83-0d05-4045-a3ad-982f971dbd5f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.300034] env[65522]: DEBUG nova.network.neutron [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Successfully created port: 0ef0d28e-583d-479c-b4f1-782fe7ad45f2 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1073.328078] env[65522]: DEBUG oslo_vmware.api [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114712, 'name': PowerOnVM_Task, 'duration_secs': 0.732671} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.328393] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1073.396086] env[65522]: DEBUG nova.compute.manager [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1073.416196] env[65522]: WARNING neutronclient.v2_0.client [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1073.417073] env[65522]: WARNING openstack [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1073.417499] env[65522]: WARNING openstack [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1073.449381] env[65522]: DEBUG oslo_vmware.api [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114695, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.508554] env[65522]: DEBUG nova.network.neutron [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1073.533709] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "07aca0b7-dea4-48b3-a391-58e0e611bfca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1073.534061] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "07aca0b7-dea4-48b3-a391-58e0e611bfca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1073.547987] env[65522]: DEBUG nova.compute.manager [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1073.549272] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5312a9d3-7d3a-4b2c-bc59-af63e3c8e998 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.684250] env[65522]: DEBUG nova.network.neutron [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1073.731090] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "99353848-2f0f-4388-9fcd-91e799342386" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.018s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1073.733802] env[65522]: DEBUG nova.scheduler.client.report [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1073.758864] env[65522]: INFO nova.compute.manager [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Took 25.10 seconds to build instance. [ 1073.775965] env[65522]: DEBUG nova.compute.manager [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1073.776696] env[65522]: DEBUG nova.virt.hardware [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1073.777404] env[65522]: DEBUG nova.virt.hardware [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1073.777404] env[65522]: DEBUG nova.virt.hardware [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1073.777404] env[65522]: DEBUG nova.virt.hardware [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1073.777648] env[65522]: DEBUG nova.virt.hardware [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1073.777648] env[65522]: DEBUG nova.virt.hardware [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1073.777804] env[65522]: DEBUG nova.virt.hardware [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1073.777958] env[65522]: DEBUG nova.virt.hardware [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1073.778178] env[65522]: DEBUG nova.virt.hardware [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1073.778354] env[65522]: DEBUG nova.virt.hardware [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1073.778523] env[65522]: DEBUG nova.virt.hardware [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1073.779787] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c75f9e-5350-4bf8-9b0a-f4ff2680ca4b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.790043] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c090829b-918a-4fbb-bbae-6262b5067049 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.950990] env[65522]: DEBUG oslo_vmware.api [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114695, 'name': ReconfigVM_Task, 'duration_secs': 5.795411} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.951584] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1073.951584] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Reconfigured VM to detach interface {{(pid=65522) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1073.952110] env[65522]: WARNING neutronclient.v2_0.client [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1073.952609] env[65522]: WARNING neutronclient.v2_0.client [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1073.953421] env[65522]: WARNING openstack [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1073.954010] env[65522]: WARNING openstack [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1074.029851] env[65522]: WARNING neutronclient.v2_0.client [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1074.037175] env[65522]: DEBUG nova.compute.manager [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1074.074131] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c509511-e7d9-46cc-9de8-cd139722e17d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 38.621s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.187084] env[65522]: DEBUG oslo_concurrency.lockutils [req-c7b2a47f-1940-488e-a0c0-8325dd4094d5 req-4da87c46-d27f-4ee1-a858-ec833b364c8f service nova] Releasing lock "refresh_cache-f6980679-bc2e-442f-b424-8cf80de86c02" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1074.187487] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Acquired lock "refresh_cache-f6980679-bc2e-442f-b424-8cf80de86c02" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1074.187677] env[65522]: DEBUG nova.network.neutron [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1074.261332] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40a5e665-93de-4a16-975b-aed78c97c882 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.611s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.409780] env[65522]: DEBUG nova.compute.manager [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1074.442036] env[65522]: DEBUG nova.virt.hardware [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1074.442036] env[65522]: DEBUG nova.virt.hardware [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1074.442036] env[65522]: DEBUG nova.virt.hardware [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1074.443195] env[65522]: DEBUG nova.virt.hardware [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1074.443195] env[65522]: DEBUG nova.virt.hardware [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1074.443195] env[65522]: DEBUG nova.virt.hardware [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1074.443195] env[65522]: DEBUG nova.virt.hardware [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1074.443459] env[65522]: DEBUG nova.virt.hardware [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1074.444027] env[65522]: DEBUG nova.virt.hardware [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1074.444027] env[65522]: DEBUG nova.virt.hardware [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1074.444027] env[65522]: DEBUG nova.virt.hardware [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1074.445253] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc018c0-a793-425c-8137-021bf5e2c30b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.455383] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-786ce2ce-8f66-4ce6-ae1e-bb1389f79ba4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.563986] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1074.690793] env[65522]: WARNING openstack [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1074.691415] env[65522]: WARNING openstack [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1074.731190] env[65522]: DEBUG nova.network.neutron [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1074.749306] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.362s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.751966] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.407s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.752286] env[65522]: DEBUG nova.objects.instance [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lazy-loading 'resources' on Instance uuid f1ce2d19-347a-4830-8501-8cceb87d4bf8 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1074.782534] env[65522]: WARNING openstack [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1074.782534] env[65522]: WARNING openstack [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1074.800038] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "99353848-2f0f-4388-9fcd-91e799342386" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1074.800038] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "99353848-2f0f-4388-9fcd-91e799342386" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.800309] env[65522]: INFO nova.compute.manager [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Attaching volume 3c40ed31-83a3-4a51-ace6-fae3fc2c638c to /dev/sdb [ 1074.845970] env[65522]: DEBUG nova.compute.manager [req-dab62adf-ddf2-43cf-a438-b0c3164028c0 req-902d223b-3b7d-4793-ba19-665bb344d00c service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Received event network-vif-plugged-0ef0d28e-583d-479c-b4f1-782fe7ad45f2 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1074.846296] env[65522]: DEBUG oslo_concurrency.lockutils [req-dab62adf-ddf2-43cf-a438-b0c3164028c0 req-902d223b-3b7d-4793-ba19-665bb344d00c service nova] Acquiring lock "821a5847-c255-4846-9041-2a0144f4539b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1074.847054] env[65522]: DEBUG oslo_concurrency.lockutils [req-dab62adf-ddf2-43cf-a438-b0c3164028c0 req-902d223b-3b7d-4793-ba19-665bb344d00c service nova] Lock "821a5847-c255-4846-9041-2a0144f4539b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.847054] env[65522]: DEBUG oslo_concurrency.lockutils [req-dab62adf-ddf2-43cf-a438-b0c3164028c0 req-902d223b-3b7d-4793-ba19-665bb344d00c service nova] Lock "821a5847-c255-4846-9041-2a0144f4539b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.847269] env[65522]: DEBUG nova.compute.manager [req-dab62adf-ddf2-43cf-a438-b0c3164028c0 req-902d223b-3b7d-4793-ba19-665bb344d00c service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] No waiting events found dispatching network-vif-plugged-0ef0d28e-583d-479c-b4f1-782fe7ad45f2 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1074.847416] env[65522]: WARNING nova.compute.manager [req-dab62adf-ddf2-43cf-a438-b0c3164028c0 req-902d223b-3b7d-4793-ba19-665bb344d00c service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Received unexpected event network-vif-plugged-0ef0d28e-583d-479c-b4f1-782fe7ad45f2 for instance with vm_state building and task_state spawning. [ 1074.853967] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a367ae97-4006-4ba7-a440-a13b7e23ce60 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.863871] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d110e24-12a2-433a-8969-f53f6e3a5566 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.882858] env[65522]: DEBUG nova.virt.block_device [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Updating existing volume attachment record: a3cae077-1960-4667-b2e7-4946aaf51c8d {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1074.886196] env[65522]: WARNING neutronclient.v2_0.client [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1074.886822] env[65522]: WARNING openstack [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1074.887184] env[65522]: WARNING openstack [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1074.943517] env[65522]: DEBUG nova.network.neutron [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Successfully updated port: 0ef0d28e-583d-479c-b4f1-782fe7ad45f2 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1075.035581] env[65522]: DEBUG nova.network.neutron [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Updating instance_info_cache with network_info: [{"id": "01011dd2-f801-46a0-8528-8305d2c08ee9", "address": "fa:16:3e:fc:1a:a1", "network": {"id": "5241c01d-364b-4796-8158-e11af4aaa9c2", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1355365557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4c0ca0038354ab48c5b2cdb004d549e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01011dd2-f8", "ovs_interfaceid": "01011dd2-f801-46a0-8528-8305d2c08ee9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1075.326320] env[65522]: INFO nova.scheduler.client.report [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleted allocation for migration 5c0d4fa1-1069-4eb5-8c35-44a984e954e4 [ 1075.326320] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1075.326320] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1075.326320] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1075.327029] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1075.327731] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1075.334422] env[65522]: INFO nova.compute.manager [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Terminating instance [ 1075.447622] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "refresh_cache-821a5847-c255-4846-9041-2a0144f4539b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.447773] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "refresh_cache-821a5847-c255-4846-9041-2a0144f4539b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1075.447928] env[65522]: DEBUG nova.network.neutron [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1075.527714] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.527924] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1075.528196] env[65522]: DEBUG nova.network.neutron [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1075.543299] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Releasing lock "refresh_cache-f6980679-bc2e-442f-b424-8cf80de86c02" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1075.543299] env[65522]: DEBUG nova.compute.manager [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Instance network_info: |[{"id": "01011dd2-f801-46a0-8528-8305d2c08ee9", "address": "fa:16:3e:fc:1a:a1", "network": {"id": "5241c01d-364b-4796-8158-e11af4aaa9c2", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1355365557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4c0ca0038354ab48c5b2cdb004d549e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01011dd2-f8", "ovs_interfaceid": "01011dd2-f801-46a0-8528-8305d2c08ee9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1075.543970] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:1a:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '40859343-2baa-45fd-88e3-ebf8aaed2b19', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01011dd2-f801-46a0-8528-8305d2c08ee9', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1075.553423] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Creating folder: Project (c4c0ca0038354ab48c5b2cdb004d549e). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1075.554433] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4f6fe460-3e6d-4ee3-9d9c-8f01fa65ef23 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.578667] env[65522]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1075.579169] env[65522]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=65522) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1075.579656] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Folder already exists: Project (c4c0ca0038354ab48c5b2cdb004d549e). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 1075.579912] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Creating folder: Instances. Parent ref: group-v994907. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1075.580219] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f7fe14b8-1718-4c94-bb76-011922ad7e18 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.597026] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Created folder: Instances in parent group-v994907. [ 1075.597445] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1075.608027] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1075.608027] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-222104f3-8590-46d5-8b68-19c0c2e3381d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.630131] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1075.630131] env[65522]: value = "task-5114718" [ 1075.630131] env[65522]: _type = "Task" [ 1075.630131] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.644239] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114718, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.707936] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc25da6a-b19b-411c-89b8-35ae8d58d61a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.718129] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17aba52c-7eb0-4d99-a7e2-6bbccea9b61e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.608800] env[65522]: DEBUG nova.compute.manager [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1076.609295] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1076.610526] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5a4387e-b3aa-4920-bec3-a2184ef09839 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "7c6f8218-602d-44f3-8012-de5a96972785" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.292s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1076.611614] env[65522]: WARNING openstack [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.611994] env[65522]: WARNING openstack [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.619657] env[65522]: WARNING neutronclient.v2_0.client [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1076.620261] env[65522]: WARNING openstack [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.620539] env[65522]: WARNING openstack [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.627479] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "7c6f8218-602d-44f3-8012-de5a96972785" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1076.628428] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "7c6f8218-602d-44f3-8012-de5a96972785" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1076.628428] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "7c6f8218-602d-44f3-8012-de5a96972785-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1076.628428] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "7c6f8218-602d-44f3-8012-de5a96972785-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1076.628428] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "7c6f8218-602d-44f3-8012-de5a96972785-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1076.632810] env[65522]: DEBUG nova.compute.manager [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Received event network-changed-2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1076.633139] env[65522]: DEBUG nova.compute.manager [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Refreshing instance network info cache due to event network-changed-2b2de45d-4c08-4bb4-810a-2f26266cbc20. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1076.633229] env[65522]: DEBUG oslo_concurrency.lockutils [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] Acquiring lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.634420] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3cde9d-0ab8-4fca-8bb6-700443605d97 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.643177] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf070496-5691-4ef2-ac34-31d4ba573aab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.646091] env[65522]: INFO nova.compute.manager [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Terminating instance [ 1076.657960] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114718, 'name': CreateVM_Task, 'duration_secs': 0.430047} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.661027] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1076.661488] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5463282f-02e9-4925-b2b8-a008f9c2052a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.665831] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1076.666896] env[65522]: WARNING neutronclient.v2_0.client [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1076.667351] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': True, 'boot_index': 0, 'mount_device': '/dev/sda', 'disk_bus': None, 'guest_format': None, 'device_type': None, 'attachment_id': 'a4e2eeb5-7784-44fc-8393-7219b63b229d', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994915', 'volume_id': 'b4d96581-08cb-4fb8-98a2-bc1957d04206', 'name': 'volume-b4d96581-08cb-4fb8-98a2-bc1957d04206', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f6980679-bc2e-442f-b424-8cf80de86c02', 'attached_at': '', 'detached_at': '', 'volume_id': 'b4d96581-08cb-4fb8-98a2-bc1957d04206', 'serial': 'b4d96581-08cb-4fb8-98a2-bc1957d04206'}, 'volume_type': None}], 'swap': None} {{(pid=65522) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1076.667455] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Root volume attach. Driver type: vmdk {{(pid=65522) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1076.671194] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7aa9a4be-b4e3-4306-8c9c-644ca5e03a4a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.672081] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-965b642d-3d0a-453d-b573-f8a6d06d398a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.677058] env[65522]: DEBUG nova.network.neutron [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1076.693275] env[65522]: DEBUG nova.compute.provider_tree [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1076.698653] env[65522]: DEBUG oslo_vmware.rw_handles [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d89eb7-c071-733d-7fc7-1639d03c9018/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1076.699636] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe3ff08-2571-44d4-8e29-18b5faf494b9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.702587] env[65522]: DEBUG oslo_vmware.api [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1076.702587] env[65522]: value = "task-5114719" [ 1076.702587] env[65522]: _type = "Task" [ 1076.702587] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.703518] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186f077c-15f4-4086-9787-3c5d93f87746 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.707720] env[65522]: WARNING openstack [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.708116] env[65522]: WARNING openstack [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.724700] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63ea811-6d41-4131-8aed-d2ae392cc9c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.728413] env[65522]: DEBUG oslo_vmware.rw_handles [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d89eb7-c071-733d-7fc7-1639d03c9018/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1076.728413] env[65522]: ERROR oslo_vmware.rw_handles [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d89eb7-c071-733d-7fc7-1639d03c9018/disk-0.vmdk due to incomplete transfer. [ 1076.731862] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-63527377-caf1-49da-86a1-7923ce87c621 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.733603] env[65522]: DEBUG oslo_vmware.api [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114719, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.739305] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-14204772-3199-484f-94c0-150a57f92845 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.747204] env[65522]: DEBUG oslo_vmware.rw_handles [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d89eb7-c071-733d-7fc7-1639d03c9018/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1076.747425] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Uploaded image c64abcc1-1c3f-4b83-a6c1-5a8864327b3b to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1076.750257] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1076.755259] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8969f651-c360-40d1-bdd5-8013697b98c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.760804] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Waiting for the task: (returnval){ [ 1076.760804] env[65522]: value = "task-5114720" [ 1076.760804] env[65522]: _type = "Task" [ 1076.760804] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.765502] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1076.765502] env[65522]: value = "task-5114721" [ 1076.765502] env[65522]: _type = "Task" [ 1076.765502] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.773967] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114720, 'name': RelocateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.780378] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114721, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.810048] env[65522]: WARNING neutronclient.v2_0.client [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1076.810945] env[65522]: WARNING openstack [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.811580] env[65522]: WARNING openstack [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.826376] env[65522]: WARNING openstack [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.826831] env[65522]: WARNING openstack [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.906785] env[65522]: WARNING neutronclient.v2_0.client [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1076.907447] env[65522]: WARNING openstack [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.907787] env[65522]: WARNING openstack [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.917899] env[65522]: DEBUG nova.network.neutron [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Updating instance_info_cache with network_info: [{"id": "0ef0d28e-583d-479c-b4f1-782fe7ad45f2", "address": "fa:16:3e:b0:ef:01", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ef0d28e-58", "ovs_interfaceid": "0ef0d28e-583d-479c-b4f1-782fe7ad45f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1077.002424] env[65522]: INFO nova.network.neutron [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Port b6097e79-9e37-4e69-b0f3-40ddbe12d95a from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1077.002857] env[65522]: DEBUG nova.network.neutron [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updating instance_info_cache with network_info: [{"id": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "address": "fa:16:3e:e9:bd:f6", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b2de45d-4c", "ovs_interfaceid": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1077.005958] env[65522]: DEBUG nova.compute.manager [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Received event network-changed-0ef0d28e-583d-479c-b4f1-782fe7ad45f2 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1077.006252] env[65522]: DEBUG nova.compute.manager [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Refreshing instance network info cache due to event network-changed-0ef0d28e-583d-479c-b4f1-782fe7ad45f2. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1077.006486] env[65522]: DEBUG oslo_concurrency.lockutils [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] Acquiring lock "refresh_cache-821a5847-c255-4846-9041-2a0144f4539b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.153304] env[65522]: DEBUG nova.compute.manager [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1077.153585] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1077.155189] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb402e4e-68c6-4604-bdff-c924a7180ce1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.169015] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1077.169343] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-abf1f841-5d76-4d6d-8a08-b0b415794aa5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.182212] env[65522]: DEBUG oslo_vmware.api [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1077.182212] env[65522]: value = "task-5114722" [ 1077.182212] env[65522]: _type = "Task" [ 1077.182212] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.198667] env[65522]: DEBUG oslo_vmware.api [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114722, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.208083] env[65522]: DEBUG nova.scheduler.client.report [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1077.230217] env[65522]: DEBUG oslo_vmware.api [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114719, 'name': PowerOffVM_Task, 'duration_secs': 0.341004} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.230523] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1077.230698] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1077.230979] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c89329c0-c8eb-423e-8fb3-240d05e3041f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.278712] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114720, 'name': RelocateVM_Task} progress is 40%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.282225] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114721, 'name': Destroy_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.284911] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "interface-95f0772d-6d3e-4e37-a0d2-9ab9de63d124-b6097e79-9e37-4e69-b0f3-40ddbe12d95a" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1077.285203] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-95f0772d-6d3e-4e37-a0d2-9ab9de63d124-b6097e79-9e37-4e69-b0f3-40ddbe12d95a" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1077.285566] env[65522]: DEBUG nova.objects.instance [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'flavor' on Instance uuid 95f0772d-6d3e-4e37-a0d2-9ab9de63d124 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1077.378650] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1077.379062] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1077.379300] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Deleting the datastore file [datastore2] c071e816-fe74-4aae-9f8e-7a1c69431f9d {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1077.379812] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-839988b0-cc2b-49e1-b148-05d08852dbe3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.394123] env[65522]: DEBUG oslo_vmware.api [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for the task: (returnval){ [ 1077.394123] env[65522]: value = "task-5114724" [ 1077.394123] env[65522]: _type = "Task" [ 1077.394123] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.408105] env[65522]: DEBUG oslo_vmware.api [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114724, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.422146] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "refresh_cache-821a5847-c255-4846-9041-2a0144f4539b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.423029] env[65522]: DEBUG nova.compute.manager [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Instance network_info: |[{"id": "0ef0d28e-583d-479c-b4f1-782fe7ad45f2", "address": "fa:16:3e:b0:ef:01", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ef0d28e-58", "ovs_interfaceid": "0ef0d28e-583d-479c-b4f1-782fe7ad45f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1077.423793] env[65522]: DEBUG oslo_concurrency.lockutils [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] Acquired lock "refresh_cache-821a5847-c255-4846-9041-2a0144f4539b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.425370] env[65522]: DEBUG nova.network.neutron [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Refreshing network info cache for port 0ef0d28e-583d-479c-b4f1-782fe7ad45f2 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1077.426060] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:ef:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0ef0d28e-583d-479c-b4f1-782fe7ad45f2', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1077.434466] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1077.435455] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1077.436321] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5227c345-c830-43ac-b9b7-de3cb94853eb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.462935] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1077.462935] env[65522]: value = "task-5114725" [ 1077.462935] env[65522]: _type = "Task" [ 1077.462935] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.476329] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114725, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.508030] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.510738] env[65522]: DEBUG oslo_concurrency.lockutils [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] Acquired lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.510933] env[65522]: DEBUG nova.network.neutron [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Refreshing network info cache for port 2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1077.697980] env[65522]: DEBUG oslo_vmware.api [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114722, 'name': PowerOffVM_Task, 'duration_secs': 0.476988} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.698466] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1077.698653] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1077.698952] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e480f18-06d2-41ba-abbb-0aa116ce565b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.714397] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.962s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1077.718072] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.154s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1077.719289] env[65522]: INFO nova.compute.claims [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1077.751878] env[65522]: INFO nova.scheduler.client.report [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleted allocations for instance f1ce2d19-347a-4830-8501-8cceb87d4bf8 [ 1077.778673] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114720, 'name': RelocateVM_Task} progress is 53%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.783025] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114721, 'name': Destroy_Task, 'duration_secs': 0.977212} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.783318] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Destroyed the VM [ 1077.783594] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1077.783880] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-930d4a61-9e6f-4feb-8c65-ae04b3f6ca06 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.790375] env[65522]: WARNING neutronclient.v2_0.client [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.791291] env[65522]: WARNING openstack [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.791763] env[65522]: WARNING openstack [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.803181] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1077.803181] env[65522]: value = "task-5114728" [ 1077.803181] env[65522]: _type = "Task" [ 1077.803181] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.803564] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1077.803797] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1077.803994] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleting the datastore file [datastore2] 7c6f8218-602d-44f3-8012-de5a96972785 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1077.804370] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8bafdc7d-bc64-4bd8-b4aa-819b4e347c31 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.821798] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114728, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.823455] env[65522]: DEBUG oslo_vmware.api [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1077.823455] env[65522]: value = "task-5114729" [ 1077.823455] env[65522]: _type = "Task" [ 1077.823455] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.839178] env[65522]: DEBUG oslo_vmware.api [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114729, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.908268] env[65522]: DEBUG oslo_vmware.api [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114724, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.922973] env[65522]: DEBUG nova.objects.instance [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'pci_requests' on Instance uuid 95f0772d-6d3e-4e37-a0d2-9ab9de63d124 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1077.936881] env[65522]: WARNING neutronclient.v2_0.client [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.937706] env[65522]: WARNING openstack [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.938291] env[65522]: WARNING openstack [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.981173] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114725, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.014532] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fdc0ad4e-0f00-4ba2-a2a9-48c2cd6bb4ad tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-b6097e79-9e37-4e69-b0f3-40ddbe12d95a" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.672s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1078.016393] env[65522]: WARNING neutronclient.v2_0.client [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.020823] env[65522]: WARNING openstack [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.020823] env[65522]: WARNING openstack [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.057809] env[65522]: WARNING openstack [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.058211] env[65522]: WARNING openstack [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.153029] env[65522]: WARNING openstack [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.153029] env[65522]: WARNING openstack [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.161204] env[65522]: WARNING neutronclient.v2_0.client [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.161769] env[65522]: WARNING openstack [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.162204] env[65522]: WARNING openstack [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.233401] env[65522]: WARNING neutronclient.v2_0.client [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.234700] env[65522]: WARNING openstack [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.234700] env[65522]: WARNING openstack [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.269999] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bd277bd9-3ed5-4b14-9904-4829c3e05dfa tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "f1ce2d19-347a-4830-8501-8cceb87d4bf8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.987s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1078.274103] env[65522]: DEBUG nova.network.neutron [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Updated VIF entry in instance network info cache for port 0ef0d28e-583d-479c-b4f1-782fe7ad45f2. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1078.274587] env[65522]: DEBUG nova.network.neutron [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Updating instance_info_cache with network_info: [{"id": "0ef0d28e-583d-479c-b4f1-782fe7ad45f2", "address": "fa:16:3e:b0:ef:01", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ef0d28e-58", "ovs_interfaceid": "0ef0d28e-583d-479c-b4f1-782fe7ad45f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1078.281878] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114720, 'name': RelocateVM_Task} progress is 65%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.319529] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114728, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.338020] env[65522]: DEBUG oslo_vmware.api [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114729, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.340958] env[65522]: DEBUG nova.network.neutron [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updated VIF entry in instance network info cache for port 2b2de45d-4c08-4bb4-810a-2f26266cbc20. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1078.341387] env[65522]: DEBUG nova.network.neutron [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updating instance_info_cache with network_info: [{"id": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "address": "fa:16:3e:e9:bd:f6", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b2de45d-4c", "ovs_interfaceid": "2b2de45d-4c08-4bb4-810a-2f26266cbc20", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1078.413874] env[65522]: DEBUG oslo_vmware.api [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114724, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.427101] env[65522]: DEBUG nova.objects.base [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Object Instance<95f0772d-6d3e-4e37-a0d2-9ab9de63d124> lazy-loaded attributes: flavor,pci_requests {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1078.427101] env[65522]: DEBUG nova.network.neutron [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1078.427484] env[65522]: WARNING neutronclient.v2_0.client [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.427987] env[65522]: WARNING neutronclient.v2_0.client [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.428979] env[65522]: WARNING openstack [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.429532] env[65522]: WARNING openstack [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.603182] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114725, 'name': CreateVM_Task, 'duration_secs': 0.662228} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.603182] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1078.603182] env[65522]: WARNING neutronclient.v2_0.client [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.603182] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.603182] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1078.603182] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1078.603182] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4a6fa58-b6c9-4d64-aa53-3b2d2cebe14a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.603182] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1078.603182] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523acd2e-c384-5eb7-1367-fcc589de5390" [ 1078.603182] env[65522]: _type = "Task" [ 1078.603182] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.603182] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523acd2e-c384-5eb7-1367-fcc589de5390, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.603182] env[65522]: DEBUG nova.policy [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b491939798e4481fb433ffb81f366b25', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '068c2387de8c406194d9b1762c7292a7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1078.777963] env[65522]: DEBUG oslo_concurrency.lockutils [req-b1b3c01e-2ca6-43b0-84ff-966293291bef req-14608d08-e5c8-40ce-a1d3-e91618060d65 service nova] Releasing lock "refresh_cache-821a5847-c255-4846-9041-2a0144f4539b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.787797] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114720, 'name': RelocateVM_Task} progress is 78%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.821547] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114728, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.837452] env[65522]: DEBUG oslo_vmware.api [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114729, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.841165] env[65522]: DEBUG nova.compute.manager [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Received event network-changed-d443e903-a2fa-4c64-b185-81fe54af293a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1078.841494] env[65522]: DEBUG nova.compute.manager [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Refreshing instance network info cache due to event network-changed-d443e903-a2fa-4c64-b185-81fe54af293a. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1078.841787] env[65522]: DEBUG oslo_concurrency.lockutils [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] Acquiring lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.841943] env[65522]: DEBUG oslo_concurrency.lockutils [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] Acquired lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1078.842137] env[65522]: DEBUG nova.network.neutron [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Refreshing network info cache for port d443e903-a2fa-4c64-b185-81fe54af293a {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1078.847478] env[65522]: DEBUG oslo_concurrency.lockutils [req-cbddf29e-5af6-4341-980d-bf4fe3ea689d req-4c9ce41a-5550-4fed-9431-9d6de316ecc6 service nova] Releasing lock "refresh_cache-11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.910826] env[65522]: DEBUG oslo_vmware.api [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114724, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.970515] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be19af55-ca2c-446f-9680-76ff5798456d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.983696] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54bd211-db13-4bc9-8d7a-44dee5fba529 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.025265] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d2f3c6-ffd8-4885-bfe6-b8cc9f943b36 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.033099] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523acd2e-c384-5eb7-1367-fcc589de5390, 'name': SearchDatastore_Task, 'duration_secs': 0.057055} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.034049] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1079.034447] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1079.034701] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.034848] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1079.035080] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1079.037125] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7d250fcb-67f2-43f8-99fa-4ce4d552198f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.042463] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4611467-9ebb-4611-a857-4ffae1fd4d4e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.060155] env[65522]: DEBUG nova.compute.provider_tree [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.062748] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1079.062947] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1079.063947] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8dba950b-9fc4-4fe3-92ac-577910164a2b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.077088] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1079.077088] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521d2997-5160-3188-b080-0ccbcead7bcf" [ 1079.077088] env[65522]: _type = "Task" [ 1079.077088] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.084910] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521d2997-5160-3188-b080-0ccbcead7bcf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.286740] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114720, 'name': RelocateVM_Task} progress is 92%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.318125] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114728, 'name': RemoveSnapshot_Task} progress is 48%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.335916] env[65522]: DEBUG oslo_vmware.api [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114729, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.349660] env[65522]: WARNING neutronclient.v2_0.client [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1079.350963] env[65522]: WARNING openstack [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1079.351649] env[65522]: WARNING openstack [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1079.409157] env[65522]: DEBUG oslo_vmware.api [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Task: {'id': task-5114724, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.965626} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.409910] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1079.410129] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1079.410325] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1079.410508] env[65522]: INFO nova.compute.manager [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Took 2.80 seconds to destroy the instance on the hypervisor. [ 1079.410760] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1079.410979] env[65522]: DEBUG nova.compute.manager [-] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1079.411077] env[65522]: DEBUG nova.network.neutron [-] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1079.411348] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1079.411937] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1079.412206] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1079.492821] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1079.565179] env[65522]: DEBUG nova.scheduler.client.report [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1079.585928] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521d2997-5160-3188-b080-0ccbcead7bcf, 'name': SearchDatastore_Task, 'duration_secs': 0.034583} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.587472] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af6964f5-c9e1-4f1b-8132-b64c453186de {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.594578] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1079.594578] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52eb6574-dd8d-b2c1-10f7-48585ff6593a" [ 1079.594578] env[65522]: _type = "Task" [ 1079.594578] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.604846] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52eb6574-dd8d-b2c1-10f7-48585ff6593a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.779263] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "8859b051-8f75-4aad-b789-42662019d4c5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1079.779608] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "8859b051-8f75-4aad-b789-42662019d4c5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1079.779870] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "8859b051-8f75-4aad-b789-42662019d4c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1079.780180] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "8859b051-8f75-4aad-b789-42662019d4c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1079.780256] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "8859b051-8f75-4aad-b789-42662019d4c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1079.781911] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114720, 'name': RelocateVM_Task} progress is 97%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.782766] env[65522]: INFO nova.compute.manager [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Terminating instance [ 1079.819032] env[65522]: DEBUG oslo_vmware.api [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114728, 'name': RemoveSnapshot_Task, 'duration_secs': 1.889702} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.819330] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1079.819594] env[65522]: INFO nova.compute.manager [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Took 27.58 seconds to snapshot the instance on the hypervisor. [ 1079.835318] env[65522]: DEBUG oslo_vmware.api [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114729, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.58951} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.835584] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1079.835780] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1079.835970] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1079.836167] env[65522]: INFO nova.compute.manager [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Took 2.68 seconds to destroy the instance on the hypervisor. [ 1079.836440] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1079.836644] env[65522]: DEBUG nova.compute.manager [-] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1079.836755] env[65522]: DEBUG nova.network.neutron [-] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1079.837023] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1079.837598] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1079.837866] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1079.855772] env[65522]: WARNING openstack [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1079.856195] env[65522]: WARNING openstack [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1079.945110] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1079.945645] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994920', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'name': 'volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '99353848-2f0f-4388-9fcd-91e799342386', 'attached_at': '', 'detached_at': '', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'serial': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1079.946750] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f291f12-1319-4d7a-84f1-6352b245e092 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.967451] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33eacb11-2eb3-4b32-a0a0-416ca7770a6f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.995958] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c/volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1079.996689] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4f10588-c706-4d4d-bc82-69c9a0806ee7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.021796] env[65522]: DEBUG oslo_vmware.api [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1080.021796] env[65522]: value = "task-5114730" [ 1080.021796] env[65522]: _type = "Task" [ 1080.021796] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.031994] env[65522]: DEBUG oslo_vmware.api [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114730, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.072465] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.355s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1080.073333] env[65522]: DEBUG nova.compute.manager [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1080.106560] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52eb6574-dd8d-b2c1-10f7-48585ff6593a, 'name': SearchDatastore_Task, 'duration_secs': 0.011948} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.106912] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1080.107361] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 821a5847-c255-4846-9041-2a0144f4539b/821a5847-c255-4846-9041-2a0144f4539b.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1080.107710] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5f9f9c7c-1328-4cec-9f7b-d6ef3ebee3f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.118074] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1080.118074] env[65522]: value = "task-5114731" [ 1080.118074] env[65522]: _type = "Task" [ 1080.118074] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.128624] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114731, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.151911] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1080.221836] env[65522]: DEBUG nova.network.neutron [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Successfully updated port: b6097e79-9e37-4e69-b0f3-40ddbe12d95a {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1080.253745] env[65522]: WARNING neutronclient.v2_0.client [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1080.254551] env[65522]: WARNING openstack [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.255014] env[65522]: WARNING openstack [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.280300] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114720, 'name': RelocateVM_Task} progress is 98%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.286338] env[65522]: DEBUG nova.compute.manager [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1080.286561] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1080.287793] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d92f27-04d6-4d52-b508-6dfa154b9e75 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.296330] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1080.296728] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e28480fb-d709-4bec-a617-bbf8de0f761a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.302932] env[65522]: DEBUG oslo_vmware.api [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1080.302932] env[65522]: value = "task-5114732" [ 1080.302932] env[65522]: _type = "Task" [ 1080.302932] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.315032] env[65522]: DEBUG oslo_vmware.api [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114732, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.381207] env[65522]: DEBUG nova.network.neutron [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updated VIF entry in instance network info cache for port d443e903-a2fa-4c64-b185-81fe54af293a. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1080.382879] env[65522]: DEBUG nova.network.neutron [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updating instance_info_cache with network_info: [{"id": "d443e903-a2fa-4c64-b185-81fe54af293a", "address": "fa:16:3e:81:ed:9d", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd443e903-a2", "ovs_interfaceid": "d443e903-a2fa-4c64-b185-81fe54af293a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1080.394060] env[65522]: DEBUG nova.compute.manager [None req-5c799099-0750-4d16-9057-01108d12cc75 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Found 1 images (rotation: 2) {{(pid=65522) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 1080.455364] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1080.455651] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1080.534138] env[65522]: DEBUG oslo_vmware.api [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114730, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.578458] env[65522]: DEBUG nova.network.neutron [-] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1080.580954] env[65522]: DEBUG nova.compute.utils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1080.583501] env[65522]: DEBUG nova.compute.manager [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1080.583501] env[65522]: DEBUG nova.network.neutron [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1080.583501] env[65522]: WARNING neutronclient.v2_0.client [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1080.583501] env[65522]: WARNING neutronclient.v2_0.client [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1080.585450] env[65522]: WARNING openstack [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.585450] env[65522]: WARNING openstack [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.629718] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114731, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.634172] env[65522]: DEBUG nova.policy [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a97b2f232f7844c993dccdc386ce3b2c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9614f18e01594bf083e619c11760fe36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1080.725129] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.782786] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114720, 'name': RelocateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.818206] env[65522]: DEBUG oslo_vmware.api [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114732, 'name': PowerOffVM_Task, 'duration_secs': 0.47533} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.818994] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1080.818994] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1080.819132] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b5d40fd-4c98-40dc-ba61-0f4979e4943b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.887960] env[65522]: DEBUG oslo_concurrency.lockutils [req-ac4a795e-1811-42f9-a955-94ad23b1a281 req-b35d7599-5a66-4bee-b5d1-4466fe855dd5 service nova] Releasing lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1080.888623] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1080.888907] env[65522]: DEBUG nova.network.neutron [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1080.919105] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1080.919431] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1080.919843] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleting the datastore file [datastore1] 8859b051-8f75-4aad-b789-42662019d4c5 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1080.920608] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dfa38f6f-a1d7-4eed-8cf0-98d2194c9545 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.932173] env[65522]: DEBUG oslo_vmware.api [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1080.932173] env[65522]: value = "task-5114734" [ 1080.932173] env[65522]: _type = "Task" [ 1080.932173] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.949599] env[65522]: DEBUG oslo_vmware.api [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114734, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.966299] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1080.966299] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1080.966299] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1080.966299] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1080.966299] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1080.966299] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1080.966299] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1080.966299] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1081.035169] env[65522]: DEBUG oslo_vmware.api [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114730, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.046739] env[65522]: DEBUG nova.network.neutron [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Successfully created port: 227f5a20-0177-480c-8cec-7485e6794c6d {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1081.073262] env[65522]: DEBUG nova.compute.manager [req-e6c0cf5b-40a9-4dc7-8260-04a73234e7a0 req-6d86c0bf-ed41-4ef3-ba7a-4a0ddc9e79c6 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Received event network-vif-plugged-b6097e79-9e37-4e69-b0f3-40ddbe12d95a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1081.073262] env[65522]: DEBUG oslo_concurrency.lockutils [req-e6c0cf5b-40a9-4dc7-8260-04a73234e7a0 req-6d86c0bf-ed41-4ef3-ba7a-4a0ddc9e79c6 service nova] Acquiring lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1081.073712] env[65522]: DEBUG oslo_concurrency.lockutils [req-e6c0cf5b-40a9-4dc7-8260-04a73234e7a0 req-6d86c0bf-ed41-4ef3-ba7a-4a0ddc9e79c6 service nova] Lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.074089] env[65522]: DEBUG oslo_concurrency.lockutils [req-e6c0cf5b-40a9-4dc7-8260-04a73234e7a0 req-6d86c0bf-ed41-4ef3-ba7a-4a0ddc9e79c6 service nova] Lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.074785] env[65522]: DEBUG nova.compute.manager [req-e6c0cf5b-40a9-4dc7-8260-04a73234e7a0 req-6d86c0bf-ed41-4ef3-ba7a-4a0ddc9e79c6 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] No waiting events found dispatching network-vif-plugged-b6097e79-9e37-4e69-b0f3-40ddbe12d95a {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1081.075105] env[65522]: WARNING nova.compute.manager [req-e6c0cf5b-40a9-4dc7-8260-04a73234e7a0 req-6d86c0bf-ed41-4ef3-ba7a-4a0ddc9e79c6 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Received unexpected event network-vif-plugged-b6097e79-9e37-4e69-b0f3-40ddbe12d95a for instance with vm_state active and task_state None. [ 1081.083061] env[65522]: INFO nova.compute.manager [-] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Took 1.67 seconds to deallocate network for instance. [ 1081.093524] env[65522]: DEBUG nova.compute.manager [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1081.130437] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114731, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.866755} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.130437] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 821a5847-c255-4846-9041-2a0144f4539b/821a5847-c255-4846-9041-2a0144f4539b.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1081.130437] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1081.130437] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-61fae90f-76de-4d8e-9724-a9fa6f1533ba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.138856] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1081.138856] env[65522]: value = "task-5114735" [ 1081.138856] env[65522]: _type = "Task" [ 1081.138856] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.153022] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114735, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.164798] env[65522]: DEBUG nova.network.neutron [-] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1081.285255] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114720, 'name': RelocateVM_Task, 'duration_secs': 4.054455} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.286924] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1081.287517] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994915', 'volume_id': 'b4d96581-08cb-4fb8-98a2-bc1957d04206', 'name': 'volume-b4d96581-08cb-4fb8-98a2-bc1957d04206', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f6980679-bc2e-442f-b424-8cf80de86c02', 'attached_at': '', 'detached_at': '', 'volume_id': 'b4d96581-08cb-4fb8-98a2-bc1957d04206', 'serial': 'b4d96581-08cb-4fb8-98a2-bc1957d04206'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1081.291577] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7145b250-d996-4194-9f27-eed5a73e0d0d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.294958] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "45b4a56e-71b8-42ec-b925-a150e667b223" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1081.295348] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "45b4a56e-71b8-42ec-b925-a150e667b223" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.315167] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c946cd2e-daf0-4c0f-8f39-eaa881111ddf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.341097] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] volume-b4d96581-08cb-4fb8-98a2-bc1957d04206/volume-b4d96581-08cb-4fb8-98a2-bc1957d04206.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1081.343756] env[65522]: DEBUG nova.compute.manager [req-bd401c66-efc0-46fc-b8ca-8a7bb9f008dc req-22167a96-5528-4f16-a20e-470890de92d8 service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Received event network-vif-deleted-ddd341a0-f73c-448c-9ff4-6ef227807e03 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1081.343989] env[65522]: DEBUG nova.compute.manager [req-bd401c66-efc0-46fc-b8ca-8a7bb9f008dc req-22167a96-5528-4f16-a20e-470890de92d8 service nova] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Received event network-vif-deleted-7ba0d5d7-364b-4bea-9e55-b168ec51ad3f {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1081.344236] env[65522]: DEBUG nova.compute.manager [req-bd401c66-efc0-46fc-b8ca-8a7bb9f008dc req-22167a96-5528-4f16-a20e-470890de92d8 service nova] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Received event network-vif-deleted-cf16c93a-755c-435b-8c86-833d8ea8afbf {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1081.345707] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a415314e-b02e-4e58-8cfe-5aaad7273774 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.376188] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Waiting for the task: (returnval){ [ 1081.376188] env[65522]: value = "task-5114736" [ 1081.376188] env[65522]: _type = "Task" [ 1081.376188] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.382144] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114736, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.392911] env[65522]: WARNING openstack [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.392911] env[65522]: WARNING openstack [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.443269] env[65522]: DEBUG oslo_vmware.api [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114734, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.218213} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.443542] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1081.443716] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1081.443963] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1081.444243] env[65522]: INFO nova.compute.manager [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1081.444507] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1081.444704] env[65522]: DEBUG nova.compute.manager [-] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1081.444809] env[65522]: DEBUG nova.network.neutron [-] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1081.445070] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1081.445581] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.445833] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.469491] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1081.469800] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.469997] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.470182] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1081.471459] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa154449-bb28-47d6-bb7d-0089b592497f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.480777] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12fd9255-69b9-4dea-b03d-0ac46c1f560a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.498938] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7627a8c9-57d1-4a22-9a83-8e35827f1ac9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.502706] env[65522]: WARNING nova.network.neutron [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] 1ea0ec37-ea40-4911-b6ad-54375253cc42 already exists in list: networks containing: ['1ea0ec37-ea40-4911-b6ad-54375253cc42']. ignoring it [ 1081.511261] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2123594f-45cf-40dc-8bd1-8214daa8675f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.545954] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179249MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1081.546160] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1081.546353] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.549771] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1081.561088] env[65522]: DEBUG oslo_vmware.api [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114730, 'name': ReconfigVM_Task, 'duration_secs': 1.059596} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.561349] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Reconfigured VM instance instance-00000058 to attach disk [datastore1] volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c/volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1081.569335] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88e01ab3-87ca-48f2-8dee-bd485091ff38 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.588189] env[65522]: DEBUG oslo_vmware.api [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1081.588189] env[65522]: value = "task-5114737" [ 1081.588189] env[65522]: _type = "Task" [ 1081.588189] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.592706] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1081.603384] env[65522]: DEBUG oslo_vmware.api [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114737, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.620215] env[65522]: WARNING openstack [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.620659] env[65522]: WARNING openstack [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.651671] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114735, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111246} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.652645] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1081.653384] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d35814-1053-4e8a-97b9-9eefdda6bc3b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.668560] env[65522]: INFO nova.compute.manager [-] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Took 1.83 seconds to deallocate network for instance. [ 1081.677475] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 821a5847-c255-4846-9041-2a0144f4539b/821a5847-c255-4846-9041-2a0144f4539b.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1081.679268] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8d2ceac6-bb61-4c03-b9ee-d742d5d0afb1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.702308] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1081.702308] env[65522]: value = "task-5114738" [ 1081.702308] env[65522]: _type = "Task" [ 1081.702308] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.711772] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114738, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.799794] env[65522]: DEBUG nova.compute.manager [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1081.885230] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114736, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.933269] env[65522]: WARNING neutronclient.v2_0.client [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1081.933978] env[65522]: WARNING openstack [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.934344] env[65522]: WARNING openstack [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1082.044258] env[65522]: WARNING openstack [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1082.044674] env[65522]: WARNING openstack [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1082.099757] env[65522]: DEBUG oslo_vmware.api [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114737, 'name': ReconfigVM_Task, 'duration_secs': 0.236506} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.100086] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994920', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'name': 'volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '99353848-2f0f-4388-9fcd-91e799342386', 'attached_at': '', 'detached_at': '', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'serial': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1082.107357] env[65522]: DEBUG nova.compute.manager [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1082.137640] env[65522]: DEBUG nova.virt.hardware [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1082.138027] env[65522]: DEBUG nova.virt.hardware [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1082.138154] env[65522]: DEBUG nova.virt.hardware [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1082.138416] env[65522]: DEBUG nova.virt.hardware [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1082.138580] env[65522]: DEBUG nova.virt.hardware [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1082.138728] env[65522]: DEBUG nova.virt.hardware [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1082.138933] env[65522]: DEBUG nova.virt.hardware [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1082.139112] env[65522]: DEBUG nova.virt.hardware [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1082.139398] env[65522]: DEBUG nova.virt.hardware [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1082.139447] env[65522]: DEBUG nova.virt.hardware [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1082.139589] env[65522]: DEBUG nova.virt.hardware [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1082.140782] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-125994bb-b763-470d-af87-9b8c7a0510ae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.155922] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa3d29a-24c2-4c1c-85ae-95a1393a36f2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.198970] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1082.202053] env[65522]: WARNING neutronclient.v2_0.client [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1082.202703] env[65522]: WARNING openstack [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1082.203089] env[65522]: WARNING openstack [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1082.219841] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114738, 'name': ReconfigVM_Task, 'duration_secs': 0.358789} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.220134] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 821a5847-c255-4846-9041-2a0144f4539b/821a5847-c255-4846-9041-2a0144f4539b.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1082.220787] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-47252cc0-913a-4fd8-8024-8414ee98d3fa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.227468] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1082.227468] env[65522]: value = "task-5114739" [ 1082.227468] env[65522]: _type = "Task" [ 1082.227468] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.236517] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114739, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.311449] env[65522]: DEBUG nova.network.neutron [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updating instance_info_cache with network_info: [{"id": "d443e903-a2fa-4c64-b185-81fe54af293a", "address": "fa:16:3e:81:ed:9d", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd443e903-a2", "ovs_interfaceid": "d443e903-a2fa-4c64-b185-81fe54af293a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b6097e79-9e37-4e69-b0f3-40ddbe12d95a", "address": "fa:16:3e:1b:7b:c6", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6097e79-9e", "ovs_interfaceid": "b6097e79-9e37-4e69-b0f3-40ddbe12d95a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1082.328260] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1082.380323] env[65522]: DEBUG nova.network.neutron [-] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1082.385569] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114736, 'name': ReconfigVM_Task, 'duration_secs': 0.567521} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.385683] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Reconfigured VM instance instance-00000060 to attach disk [datastore1] volume-b4d96581-08cb-4fb8-98a2-bc1957d04206/volume-b4d96581-08cb-4fb8-98a2-bc1957d04206.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1082.392535] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d754f2d-7bef-4435-806b-201d6989eee3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.412406] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Waiting for the task: (returnval){ [ 1082.412406] env[65522]: value = "task-5114740" [ 1082.412406] env[65522]: _type = "Task" [ 1082.412406] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.423827] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114740, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.513712] env[65522]: DEBUG nova.compute.manager [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1082.514788] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7faea5-ad95-4d88-90ca-1b65f5e413b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.591259] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7617189c-a902-42e7-8165-0e7c4a1de06d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1082.591410] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1082.591529] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 4aca4e87-b923-49e4-88d8-7ff51bcbe993 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1082.591642] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 8859b051-8f75-4aad-b789-42662019d4c5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1082.592635] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 9f342b89-bde2-4c35-ae42-cfe1e6973b74 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1082.592635] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1082.592635] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 99353848-2f0f-4388-9fcd-91e799342386 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1082.592635] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 95f0772d-6d3e-4e37-a0d2-9ab9de63d124 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1082.592635] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 79dd37b5-6b30-48ab-9f00-78214cbd132d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1082.592635] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7c6f8218-602d-44f3-8012-de5a96972785 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1082.592635] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance c071e816-fe74-4aae-9f8e-7a1c69431f9d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1082.592635] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 210184da-48db-4926-ae00-7c0f29622810 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1082.592990] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance f6980679-bc2e-442f-b424-8cf80de86c02 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1082.592990] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 821a5847-c255-4846-9041-2a0144f4539b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1082.592990] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 07aca0b7-dea4-48b3-a391-58e0e611bfca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1082.664095] env[65522]: DEBUG nova.network.neutron [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Successfully updated port: 227f5a20-0177-480c-8cec-7485e6794c6d {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1082.739243] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114739, 'name': Rename_Task, 'duration_secs': 0.30264} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.739585] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1082.740267] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0c005b7d-ad4e-4088-a7f8-117b2a9cf7c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.747754] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1082.747754] env[65522]: value = "task-5114741" [ 1082.747754] env[65522]: _type = "Task" [ 1082.747754] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.760300] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114741, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.816023] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1082.816496] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.816663] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1082.817536] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a027b1-62d7-457d-b7f9-677867e64d50 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.835545] env[65522]: DEBUG nova.virt.hardware [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1082.835789] env[65522]: DEBUG nova.virt.hardware [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1082.835944] env[65522]: DEBUG nova.virt.hardware [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1082.836134] env[65522]: DEBUG nova.virt.hardware [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1082.836280] env[65522]: DEBUG nova.virt.hardware [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1082.836427] env[65522]: DEBUG nova.virt.hardware [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1082.836626] env[65522]: DEBUG nova.virt.hardware [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1082.836781] env[65522]: DEBUG nova.virt.hardware [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1082.836945] env[65522]: DEBUG nova.virt.hardware [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1082.837114] env[65522]: DEBUG nova.virt.hardware [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1082.837287] env[65522]: DEBUG nova.virt.hardware [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1082.843809] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Reconfiguring VM to attach interface {{(pid=65522) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1082.844203] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b476690-bd2d-4583-9a4e-2e8d6e988f16 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.863312] env[65522]: DEBUG oslo_vmware.api [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1082.863312] env[65522]: value = "task-5114742" [ 1082.863312] env[65522]: _type = "Task" [ 1082.863312] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.873224] env[65522]: DEBUG oslo_vmware.api [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114742, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.888090] env[65522]: INFO nova.compute.manager [-] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Took 1.44 seconds to deallocate network for instance. [ 1082.924701] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114740, 'name': ReconfigVM_Task, 'duration_secs': 0.213313} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.925433] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994915', 'volume_id': 'b4d96581-08cb-4fb8-98a2-bc1957d04206', 'name': 'volume-b4d96581-08cb-4fb8-98a2-bc1957d04206', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f6980679-bc2e-442f-b424-8cf80de86c02', 'attached_at': '', 'detached_at': '', 'volume_id': 'b4d96581-08cb-4fb8-98a2-bc1957d04206', 'serial': 'b4d96581-08cb-4fb8-98a2-bc1957d04206'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1082.926033] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-89c12d2b-bafb-4879-82e3-6a5db1417c30 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.935769] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Waiting for the task: (returnval){ [ 1082.935769] env[65522]: value = "task-5114743" [ 1082.935769] env[65522]: _type = "Task" [ 1082.935769] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.951380] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114743, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.035818] env[65522]: INFO nova.compute.manager [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] instance snapshotting [ 1083.036286] env[65522]: DEBUG nova.objects.instance [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'flavor' on Instance uuid 9f342b89-bde2-4c35-ae42-cfe1e6973b74 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1083.097133] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 45b4a56e-71b8-42ec-b925-a150e667b223 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1083.097133] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1083.097133] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3264MB phys_disk=100GB used_disk=13GB total_vcpus=48 used_vcpus=14 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '14', 'num_vm_active': '11', 'num_task_deleting': '2', 'num_os_type_None': '14', 'num_proj_fa11b46d9fe144f391233e6eb9c819d7': '1', 'io_workload': '3', 'num_task_None': '9', 'num_proj_91597233ae9c44c094f4c32d90332fa6': '2', 'num_proj_a80f979f3dc0477e9462b47f7aa87f14': '1', 'num_proj_9614f18e01594bf083e619c11760fe36': '3', 'num_proj_1546bbdca07c443d84abea1155cfde37': '1', 'num_proj_3ea0fab7de6c47eeaf3bdde413ae3901': '1', 'num_proj_0d1239b79ae94cceb89ae7a8bd57da08': '1', 'num_proj_068c2387de8c406194d9b1762c7292a7': '2', 'num_vm_building': '3', 'num_task_spawning': '2', 'num_proj_c4c0ca0038354ab48c5b2cdb004d549e': '1', 'num_proj_f93c790ff61543bd8e134bcf9cb20bb2': '1', 'num_task_block_device_mapping': '1'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1083.151255] env[65522]: DEBUG nova.objects.instance [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lazy-loading 'flavor' on Instance uuid 99353848-2f0f-4388-9fcd-91e799342386 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1083.167169] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "refresh_cache-07aca0b7-dea4-48b3-a391-58e0e611bfca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.167356] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "refresh_cache-07aca0b7-dea4-48b3-a391-58e0e611bfca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1083.167558] env[65522]: DEBUG nova.network.neutron [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1083.212305] env[65522]: DEBUG nova.compute.manager [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Received event network-changed-b6097e79-9e37-4e69-b0f3-40ddbe12d95a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1083.212305] env[65522]: DEBUG nova.compute.manager [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Refreshing instance network info cache due to event network-changed-b6097e79-9e37-4e69-b0f3-40ddbe12d95a. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1083.212305] env[65522]: DEBUG oslo_concurrency.lockutils [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] Acquiring lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.212305] env[65522]: DEBUG oslo_concurrency.lockutils [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] Acquired lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1083.212305] env[65522]: DEBUG nova.network.neutron [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Refreshing network info cache for port b6097e79-9e37-4e69-b0f3-40ddbe12d95a {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1083.263190] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114741, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.342095] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475ea702-3480-4517-8388-43726d70c026 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.352796] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26e7782-d870-473d-af22-523dd1b2e219 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.393984] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61e7484-bd44-448e-b1bb-584a1dd59141 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.397977] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1083.405203] env[65522]: DEBUG oslo_vmware.api [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114742, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.408723] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcacbead-8698-4ff2-8cfd-40b9630906b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.423999] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.447685] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114743, 'name': Rename_Task, 'duration_secs': 0.154653} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.447685] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1083.447927] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b99faef0-5a0e-449a-bc70-b0ea256dc456 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.455754] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Waiting for the task: (returnval){ [ 1083.455754] env[65522]: value = "task-5114744" [ 1083.455754] env[65522]: _type = "Task" [ 1083.455754] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.465573] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114744, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.523091] env[65522]: DEBUG nova.compute.manager [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Received event network-vif-plugged-227f5a20-0177-480c-8cec-7485e6794c6d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1083.523324] env[65522]: DEBUG oslo_concurrency.lockutils [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] Acquiring lock "07aca0b7-dea4-48b3-a391-58e0e611bfca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1083.523530] env[65522]: DEBUG oslo_concurrency.lockutils [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] Lock "07aca0b7-dea4-48b3-a391-58e0e611bfca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1083.523678] env[65522]: DEBUG oslo_concurrency.lockutils [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] Lock "07aca0b7-dea4-48b3-a391-58e0e611bfca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1083.523842] env[65522]: DEBUG nova.compute.manager [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] No waiting events found dispatching network-vif-plugged-227f5a20-0177-480c-8cec-7485e6794c6d {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1083.524017] env[65522]: WARNING nova.compute.manager [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Received unexpected event network-vif-plugged-227f5a20-0177-480c-8cec-7485e6794c6d for instance with vm_state building and task_state spawning. [ 1083.524202] env[65522]: DEBUG nova.compute.manager [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Received event network-changed-227f5a20-0177-480c-8cec-7485e6794c6d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1083.524318] env[65522]: DEBUG nova.compute.manager [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Refreshing instance network info cache due to event network-changed-227f5a20-0177-480c-8cec-7485e6794c6d. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1083.524473] env[65522]: DEBUG oslo_concurrency.lockutils [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] Acquiring lock "refresh_cache-07aca0b7-dea4-48b3-a391-58e0e611bfca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.544698] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c735fdb1-88e8-4ece-9200-1c650fe15699 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.565865] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6d8be4-a5a9-46f0-b64f-6c97727def04 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.658193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e7ce4812-3f42-4c4d-90fd-3f32f36671d1 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "99353848-2f0f-4388-9fcd-91e799342386" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.858s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1083.671254] env[65522]: WARNING openstack [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.671812] env[65522]: WARNING openstack [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.706962] env[65522]: INFO nova.compute.manager [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Rebuilding instance [ 1083.712822] env[65522]: DEBUG nova.network.neutron [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1083.715334] env[65522]: WARNING neutronclient.v2_0.client [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1083.715986] env[65522]: WARNING openstack [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.716393] env[65522]: WARNING openstack [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.743440] env[65522]: WARNING openstack [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.743995] env[65522]: WARNING openstack [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.765222] env[65522]: DEBUG nova.compute.manager [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1083.766117] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45138b1c-9004-4ac0-939c-f777da53c61d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.772716] env[65522]: DEBUG oslo_vmware.api [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114741, 'name': PowerOnVM_Task, 'duration_secs': 0.752691} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.773460] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1083.773676] env[65522]: INFO nova.compute.manager [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Took 9.37 seconds to spawn the instance on the hypervisor. [ 1083.773868] env[65522]: DEBUG nova.compute.manager [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1083.774775] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d982e0a9-71da-481e-bc19-b83ce84fc9f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.831658] env[65522]: WARNING neutronclient.v2_0.client [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1083.832225] env[65522]: WARNING openstack [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.832490] env[65522]: WARNING openstack [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.895673] env[65522]: WARNING openstack [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.896045] env[65522]: WARNING openstack [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.913792] env[65522]: DEBUG oslo_vmware.api [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114742, 'name': ReconfigVM_Task, 'duration_secs': 0.878883} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.914038] env[65522]: WARNING neutronclient.v2_0.client [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1083.914293] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1083.914501] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Reconfigured VM to attach interface {{(pid=65522) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1083.927119] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1083.947822] env[65522]: DEBUG nova.network.neutron [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Updating instance_info_cache with network_info: [{"id": "227f5a20-0177-480c-8cec-7485e6794c6d", "address": "fa:16:3e:5f:12:fb", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap227f5a20-01", "ovs_interfaceid": "227f5a20-0177-480c-8cec-7485e6794c6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1083.968321] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114744, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.970308] env[65522]: WARNING neutronclient.v2_0.client [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1083.971039] env[65522]: WARNING openstack [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.971458] env[65522]: WARNING openstack [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.058692] env[65522]: DEBUG nova.network.neutron [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updated VIF entry in instance network info cache for port b6097e79-9e37-4e69-b0f3-40ddbe12d95a. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1084.059161] env[65522]: DEBUG nova.network.neutron [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updating instance_info_cache with network_info: [{"id": "d443e903-a2fa-4c64-b185-81fe54af293a", "address": "fa:16:3e:81:ed:9d", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd443e903-a2", "ovs_interfaceid": "d443e903-a2fa-4c64-b185-81fe54af293a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b6097e79-9e37-4e69-b0f3-40ddbe12d95a", "address": "fa:16:3e:1b:7b:c6", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6097e79-9e", "ovs_interfaceid": "b6097e79-9e37-4e69-b0f3-40ddbe12d95a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1084.078012] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1084.078345] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-77e4d182-790d-4258-90b1-ecf0da13ebd8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.089114] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1084.089114] env[65522]: value = "task-5114745" [ 1084.089114] env[65522]: _type = "Task" [ 1084.089114] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.099901] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114745, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.307511] env[65522]: INFO nova.compute.manager [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Took 15.82 seconds to build instance. [ 1084.420116] env[65522]: DEBUG oslo_concurrency.lockutils [None req-15a29d5a-7756-483f-bd3e-63b1180faa7a tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-95f0772d-6d3e-4e37-a0d2-9ab9de63d124-b6097e79-9e37-4e69-b0f3-40ddbe12d95a" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.134s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.440511] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1084.440737] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.894s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.441329] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.849s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1084.441518] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.443850] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.245s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1084.444101] env[65522]: DEBUG nova.objects.instance [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lazy-loading 'resources' on Instance uuid 7c6f8218-602d-44f3-8012-de5a96972785 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1084.451696] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "refresh_cache-07aca0b7-dea4-48b3-a391-58e0e611bfca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1084.452164] env[65522]: DEBUG nova.compute.manager [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Instance network_info: |[{"id": "227f5a20-0177-480c-8cec-7485e6794c6d", "address": "fa:16:3e:5f:12:fb", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap227f5a20-01", "ovs_interfaceid": "227f5a20-0177-480c-8cec-7485e6794c6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1084.452474] env[65522]: DEBUG oslo_concurrency.lockutils [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] Acquired lock "refresh_cache-07aca0b7-dea4-48b3-a391-58e0e611bfca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1084.452676] env[65522]: DEBUG nova.network.neutron [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Refreshing network info cache for port 227f5a20-0177-480c-8cec-7485e6794c6d {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1084.453879] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5f:12:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10b81051-1eb1-406b-888c-4548c470c77e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '227f5a20-0177-480c-8cec-7485e6794c6d', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1084.461517] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1084.463161] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1084.466644] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-78d0812c-e055-42c5-885e-e04d4f1ace6d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.483927] env[65522]: INFO nova.scheduler.client.report [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Deleted allocations for instance c071e816-fe74-4aae-9f8e-7a1c69431f9d [ 1084.492026] env[65522]: DEBUG oslo_vmware.api [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114744, 'name': PowerOnVM_Task, 'duration_secs': 0.764308} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.493701] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1084.493959] env[65522]: INFO nova.compute.manager [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Took 10.72 seconds to spawn the instance on the hypervisor. [ 1084.494344] env[65522]: DEBUG nova.compute.manager [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1084.494816] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1084.494816] env[65522]: value = "task-5114746" [ 1084.494816] env[65522]: _type = "Task" [ 1084.494816] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.495633] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d39c841c-8366-4480-8878-3ff5ece1da14 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.513351] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114746, 'name': CreateVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.562427] env[65522]: DEBUG oslo_concurrency.lockutils [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] Releasing lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1084.562975] env[65522]: DEBUG nova.compute.manager [req-61c2db99-78b6-48ad-b0b1-00a5518f0b51 req-33c9892f-e84d-4a68-9301-19ef7fae7788 service nova] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Received event network-vif-deleted-cc030045-9833-4322-9675-37b3f0f31161 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1084.600149] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114745, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.787982] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1084.788353] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3cfb7cac-59cd-4493-a5a4-c323e6011bdf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.796195] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1084.796195] env[65522]: value = "task-5114747" [ 1084.796195] env[65522]: _type = "Task" [ 1084.796195] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.807820] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114747, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.810425] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2ddf661-2934-4ba8-9aea-61e544866003 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "821a5847-c255-4846-9041-2a0144f4539b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.328s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.963768] env[65522]: WARNING neutronclient.v2_0.client [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1084.964271] env[65522]: WARNING openstack [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.965231] env[65522]: WARNING openstack [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.997965] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55ab7596-526b-44cb-8c1b-61c87a197da5 tempest-ServersTestMultiNic-517359145 tempest-ServersTestMultiNic-517359145-project-member] Lock "c071e816-fe74-4aae-9f8e-7a1c69431f9d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.672s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.010423] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114746, 'name': CreateVM_Task, 'duration_secs': 0.506606} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.013905] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1085.019047] env[65522]: WARNING neutronclient.v2_0.client [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1085.020435] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.020435] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1085.020435] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1085.022149] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2d1d4e8-97ad-4677-a965-618e6db57a3f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.025090] env[65522]: INFO nova.compute.manager [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Took 17.45 seconds to build instance. [ 1085.032630] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1085.032630] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d9166e-84a1-9966-8d0f-53522e97ec19" [ 1085.032630] env[65522]: _type = "Task" [ 1085.032630] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.046440] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d9166e-84a1-9966-8d0f-53522e97ec19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.105679] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114745, 'name': CreateSnapshot_Task, 'duration_secs': 0.963512} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.106143] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1085.106944] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e95c0b4-3b0d-4057-bd15-b1db74e3b027 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.128927] env[65522]: WARNING openstack [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1085.129342] env[65522]: WARNING openstack [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1085.219526] env[65522]: WARNING neutronclient.v2_0.client [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1085.219874] env[65522]: WARNING openstack [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1085.220209] env[65522]: WARNING openstack [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1085.229994] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a240d0b-6e13-4225-be63-78a5492c6d80 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.238391] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01592bf6-50ad-4b9b-895b-f54f19bc419e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.282363] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70c0396-3de6-4f9d-a7b9-7fdfadde4f8f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.292989] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5759d14-57a8-4e29-91cb-aa270419ae3d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.308614] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114747, 'name': PowerOffVM_Task, 'duration_secs': 0.353818} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.316976] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1085.319408] env[65522]: DEBUG nova.compute.provider_tree [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1085.332177] env[65522]: DEBUG nova.network.neutron [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Updated VIF entry in instance network info cache for port 227f5a20-0177-480c-8cec-7485e6794c6d. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1085.332437] env[65522]: DEBUG nova.network.neutron [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Updating instance_info_cache with network_info: [{"id": "227f5a20-0177-480c-8cec-7485e6794c6d", "address": "fa:16:3e:5f:12:fb", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap227f5a20-01", "ovs_interfaceid": "227f5a20-0177-480c-8cec-7485e6794c6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1085.437506] env[65522]: INFO nova.compute.manager [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Detaching volume 3c40ed31-83a3-4a51-ace6-fae3fc2c638c [ 1085.478851] env[65522]: INFO nova.virt.block_device [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Attempting to driver detach volume 3c40ed31-83a3-4a51-ace6-fae3fc2c638c from mountpoint /dev/sdb [ 1085.478851] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1085.478993] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994920', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'name': 'volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '99353848-2f0f-4388-9fcd-91e799342386', 'attached_at': '', 'detached_at': '', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'serial': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1085.480642] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08f8deb0-cf53-4312-8210-ec01b603d233 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.509027] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfae8d26-a22b-4926-a1db-433bc163239a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.517553] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6802f44-cb1c-42d0-aeae-5d2ddc575c42 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.541208] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7ed5b939-f599-4515-b660-fd97722713ed tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Lock "f6980679-bc2e-442f-b424-8cf80de86c02" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.980s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.545523] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a91b6e19-831e-4d58-92be-9fbcf4530c88 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.567257] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d9166e-84a1-9966-8d0f-53522e97ec19, 'name': SearchDatastore_Task, 'duration_secs': 0.015024} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.567412] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The volume has not been displaced from its original location: [datastore1] volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c/volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1085.573305] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Reconfiguring VM instance instance-00000058 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1085.573845] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1085.574730] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1085.574730] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.574730] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1085.574952] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1085.575175] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a6f5d77-2cb7-41fc-9445-79584c09f679 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.590502] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-999be83b-5465-4c88-bbfb-95965ded599f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.600157] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1085.600157] env[65522]: value = "task-5114748" [ 1085.600157] env[65522]: _type = "Task" [ 1085.600157] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.601941] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1085.601941] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1085.605338] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cbb2ef6-39f2-4b89-95ff-708c3759d11a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.616430] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114748, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.616789] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1085.616789] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52852b51-2033-510f-37da-c3b66e6f716a" [ 1085.616789] env[65522]: _type = "Task" [ 1085.616789] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.630040] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1085.634809] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9215d31b-b51b-451b-bbbb-9ee1b072990f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.638311] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52852b51-2033-510f-37da-c3b66e6f716a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.644761] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1085.644761] env[65522]: value = "task-5114749" [ 1085.644761] env[65522]: _type = "Task" [ 1085.644761] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.653994] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114749, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.823190] env[65522]: DEBUG nova.scheduler.client.report [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1085.835658] env[65522]: DEBUG oslo_concurrency.lockutils [req-787ce599-5b20-4e7a-b64b-440cf4efda0b req-29f58119-994b-403d-95e7-a08774d558d5 service nova] Releasing lock "refresh_cache-07aca0b7-dea4-48b3-a391-58e0e611bfca" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1086.117629] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114748, 'name': ReconfigVM_Task, 'duration_secs': 0.423015} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.117629] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Reconfigured VM instance instance-00000058 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1086.119342] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "821a5847-c255-4846-9041-2a0144f4539b" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1086.119342] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "821a5847-c255-4846-9041-2a0144f4539b" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1086.119512] env[65522]: INFO nova.compute.manager [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Shelving [ 1086.121510] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1726f07d-1334-4db0-b978-01c4635b4697 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.147805] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52852b51-2033-510f-37da-c3b66e6f716a, 'name': SearchDatastore_Task, 'duration_secs': 0.018713} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.153327] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1086.153327] env[65522]: value = "task-5114750" [ 1086.153327] env[65522]: _type = "Task" [ 1086.153327] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.153675] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-378254db-c370-4c02-9149-6570c63ae35c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.165497] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "interface-95f0772d-6d3e-4e37-a0d2-9ab9de63d124-b6097e79-9e37-4e69-b0f3-40ddbe12d95a" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1086.165918] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-95f0772d-6d3e-4e37-a0d2-9ab9de63d124-b6097e79-9e37-4e69-b0f3-40ddbe12d95a" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1086.167696] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114749, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.171108] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1086.171108] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c59e1c-2ac4-7789-ab72-18dd4a4a7ecf" [ 1086.171108] env[65522]: _type = "Task" [ 1086.171108] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.180063] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.187457] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c59e1c-2ac4-7789-ab72-18dd4a4a7ecf, 'name': SearchDatastore_Task, 'duration_secs': 0.01386} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.187776] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1086.188084] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 07aca0b7-dea4-48b3-a391-58e0e611bfca/07aca0b7-dea4-48b3-a391-58e0e611bfca.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1086.188406] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40024605-b5c9-4aa8-9464-f7c8edea0c73 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.196470] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1086.196470] env[65522]: value = "task-5114751" [ 1086.196470] env[65522]: _type = "Task" [ 1086.196470] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.206386] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114751, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.329928] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.884s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1086.331366] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.003s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1086.334117] env[65522]: INFO nova.compute.claims [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1086.362459] env[65522]: INFO nova.scheduler.client.report [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleted allocations for instance 7c6f8218-602d-44f3-8012-de5a96972785 [ 1086.447481] env[65522]: DEBUG nova.compute.manager [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Received event network-changed-01011dd2-f801-46a0-8528-8305d2c08ee9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1086.448115] env[65522]: DEBUG nova.compute.manager [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Refreshing instance network info cache due to event network-changed-01011dd2-f801-46a0-8528-8305d2c08ee9. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1086.448403] env[65522]: DEBUG oslo_concurrency.lockutils [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] Acquiring lock "refresh_cache-f6980679-bc2e-442f-b424-8cf80de86c02" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.448581] env[65522]: DEBUG oslo_concurrency.lockutils [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] Acquired lock "refresh_cache-f6980679-bc2e-442f-b424-8cf80de86c02" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1086.448838] env[65522]: DEBUG nova.network.neutron [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Refreshing network info cache for port 01011dd2-f801-46a0-8528-8305d2c08ee9 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1086.660798] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114749, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.672404] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.672404] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1086.672404] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114750, 'name': ReconfigVM_Task, 'duration_secs': 0.175854} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.673561] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bac4b59-4912-4a33-90b0-9c326ebdc97e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.679186] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994920', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'name': 'volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '99353848-2f0f-4388-9fcd-91e799342386', 'attached_at': '', 'detached_at': '', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'serial': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1086.707488] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6584368-c037-4655-acc2-3d2197c3ba4c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.743564] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114751, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.744592] env[65522]: WARNING neutronclient.v2_0.client [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1086.750902] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Reconfiguring VM to detach interface {{(pid=65522) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1086.751382] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ad39748-32fb-46a0-a7f2-6d254bab75cc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.775366] env[65522]: DEBUG oslo_vmware.api [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1086.775366] env[65522]: value = "task-5114752" [ 1086.775366] env[65522]: _type = "Task" [ 1086.775366] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.786405] env[65522]: DEBUG oslo_vmware.api [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114752, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.871902] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b84dc884-5002-4c05-9f7a-f55dd89981db tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "7c6f8218-602d-44f3-8012-de5a96972785" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.243s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1086.952733] env[65522]: WARNING neutronclient.v2_0.client [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1086.953483] env[65522]: WARNING openstack [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.953847] env[65522]: WARNING openstack [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.147928] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1087.148445] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b15e4718-ba3c-41a6-af0b-8ca74d125c05 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.161245] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114749, 'name': CloneVM_Task} progress is 95%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.162228] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1087.162228] env[65522]: value = "task-5114753" [ 1087.162228] env[65522]: _type = "Task" [ 1087.162228] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.212522] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114751, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595896} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.212801] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 07aca0b7-dea4-48b3-a391-58e0e611bfca/07aca0b7-dea4-48b3-a391-58e0e611bfca.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1087.213138] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1087.213297] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-88431559-8f39-4f5d-881e-2a5b81e95460 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.222450] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1087.222450] env[65522]: value = "task-5114754" [ 1087.222450] env[65522]: _type = "Task" [ 1087.222450] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.232751] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114754, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.238035] env[65522]: WARNING openstack [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.238756] env[65522]: WARNING openstack [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.287452] env[65522]: DEBUG oslo_vmware.api [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.378606] env[65522]: WARNING neutronclient.v2_0.client [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1087.379713] env[65522]: WARNING openstack [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.380598] env[65522]: WARNING openstack [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.530919] env[65522]: DEBUG nova.network.neutron [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Updated VIF entry in instance network info cache for port 01011dd2-f801-46a0-8528-8305d2c08ee9. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1087.531889] env[65522]: DEBUG nova.network.neutron [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Updating instance_info_cache with network_info: [{"id": "01011dd2-f801-46a0-8528-8305d2c08ee9", "address": "fa:16:3e:fc:1a:a1", "network": {"id": "5241c01d-364b-4796-8158-e11af4aaa9c2", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1355365557-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.217", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4c0ca0038354ab48c5b2cdb004d549e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "40859343-2baa-45fd-88e3-ebf8aaed2b19", "external-id": "nsx-vlan-transportzone-10", "segmentation_id": 10, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01011dd2-f8", "ovs_interfaceid": "01011dd2-f801-46a0-8528-8305d2c08ee9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1087.580640] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc319f5-16d7-4f2e-8be5-4f9670f73772 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.591735] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a39a79f-592c-4249-9c6b-3ac09e3c57f2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.628145] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72af27a3-8239-43a8-9922-b877d254a0fe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.636760] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c71999-cd65-4d7f-a643-7b9d5159778d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.650977] env[65522]: DEBUG nova.compute.provider_tree [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1087.660827] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114749, 'name': CloneVM_Task, 'duration_secs': 1.694039} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.661926] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Created linked-clone VM from snapshot [ 1087.662508] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd9c425-f94f-40b5-80c9-bf1145676f7c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.675564] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Uploading image 9f4c00f3-d40c-4eeb-8605-5c426257ec5a {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1087.681077] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114753, 'name': PowerOffVM_Task, 'duration_secs': 0.261048} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.681346] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1087.682856] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a51ad43-dabf-4ce7-a571-5ba68c45e4a7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.705597] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5ae85f-c38a-4cb2-932f-6d5bb0ffb3e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.721523] env[65522]: DEBUG oslo_vmware.rw_handles [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1087.721523] env[65522]: value = "vm-994926" [ 1087.721523] env[65522]: _type = "VirtualMachine" [ 1087.721523] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1087.721523] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d4e47a30-ec28-4a6c-9a6c-6116ce9c812a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.729354] env[65522]: DEBUG oslo_vmware.rw_handles [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lease: (returnval){ [ 1087.729354] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52155a8c-e82e-8c51-f528-075e0543970e" [ 1087.729354] env[65522]: _type = "HttpNfcLease" [ 1087.729354] env[65522]: } obtained for exporting VM: (result){ [ 1087.729354] env[65522]: value = "vm-994926" [ 1087.729354] env[65522]: _type = "VirtualMachine" [ 1087.729354] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1087.730070] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the lease: (returnval){ [ 1087.730070] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52155a8c-e82e-8c51-f528-075e0543970e" [ 1087.730070] env[65522]: _type = "HttpNfcLease" [ 1087.730070] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1087.736829] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114754, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.149465} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.737543] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1087.738396] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22b25f7-277a-4d81-a0a6-e01a6a036083 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.742827] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1087.742827] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52155a8c-e82e-8c51-f528-075e0543970e" [ 1087.742827] env[65522]: _type = "HttpNfcLease" [ 1087.742827] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1087.743517] env[65522]: DEBUG oslo_vmware.rw_handles [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1087.743517] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52155a8c-e82e-8c51-f528-075e0543970e" [ 1087.743517] env[65522]: _type = "HttpNfcLease" [ 1087.743517] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1087.744813] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6fadd48-558c-416e-96db-b9f1ab29d85b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.757797] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1087.767191] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 07aca0b7-dea4-48b3-a391-58e0e611bfca/07aca0b7-dea4-48b3-a391-58e0e611bfca.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1087.768795] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d6594e0c-bbaa-4705-ba99-461caff169d5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.770793] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ddfeaa8c-32a6-4b98-b62f-6fd705e185f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.790549] env[65522]: DEBUG oslo_vmware.rw_handles [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525e465f-a58c-6957-b27c-f4055735341a/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1087.790549] env[65522]: DEBUG oslo_vmware.rw_handles [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525e465f-a58c-6957-b27c-f4055735341a/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1087.852850] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1087.852850] env[65522]: value = "task-5114756" [ 1087.852850] env[65522]: _type = "Task" [ 1087.852850] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.854472] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1087.854472] env[65522]: value = "task-5114757" [ 1087.854472] env[65522]: _type = "Task" [ 1087.854472] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.862552] env[65522]: DEBUG oslo_vmware.api [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.876043] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114757, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.876389] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1087.876588] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1087.876779] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994920', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'name': 'volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '99353848-2f0f-4388-9fcd-91e799342386', 'attached_at': '', 'detached_at': '', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'serial': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1087.877594] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc07372-dfaf-4cb0-a203-1de01ee6268f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.899740] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2d7c29-e588-48aa-b84d-241b8fc88be8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.905498] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-aa83778a-88dd-44a7-b377-4625009c34d7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.909377] env[65522]: WARNING nova.virt.vmwareapi.driver [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1087.909740] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1087.910947] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49215fdb-b40c-476b-9834-9352eab56a4d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.921111] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1087.921482] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a8508460-4892-4cf7-ab72-54d794628d17 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.015244] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1088.015703] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1088.016688] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleting the datastore file [datastore2] 99353848-2f0f-4388-9fcd-91e799342386 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1088.016688] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c470fe4c-ad1f-4e95-8c1c-ba69c72257cc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.025267] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1088.025267] env[65522]: value = "task-5114759" [ 1088.025267] env[65522]: _type = "Task" [ 1088.025267] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.037059] env[65522]: DEBUG oslo_concurrency.lockutils [req-ae538c37-7e1b-4324-91bf-c6948d65a7f1 req-bbba5587-63b6-470a-8721-6a7641b69453 service nova] Releasing lock "refresh_cache-f6980679-bc2e-442f-b424-8cf80de86c02" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1088.037059] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114759, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.157307] env[65522]: DEBUG nova.scheduler.client.report [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1088.222490] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1088.222824] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-de6af301-6f02-4013-8093-68ce5f50c13e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.232371] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1088.232371] env[65522]: value = "task-5114760" [ 1088.232371] env[65522]: _type = "Task" [ 1088.232371] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.245422] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114760, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.302935] env[65522]: DEBUG oslo_vmware.api [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.328382] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "cb62df50-2f86-469a-b2ee-fd3754d61c83" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1088.328796] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "cb62df50-2f86-469a-b2ee-fd3754d61c83" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1088.374885] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114757, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.536557] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114759, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.239755} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.536933] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1088.537194] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1088.537383] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1088.662983] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.332s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1088.663572] env[65522]: DEBUG nova.compute.manager [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1088.666914] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.269s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1088.666914] env[65522]: DEBUG nova.objects.instance [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lazy-loading 'resources' on Instance uuid 8859b051-8f75-4aad-b789-42662019d4c5 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.743172] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114760, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.802425] env[65522]: DEBUG oslo_vmware.api [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.833068] env[65522]: DEBUG nova.compute.manager [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1088.873234] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114757, 'name': ReconfigVM_Task, 'duration_secs': 0.822169} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.874080] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 07aca0b7-dea4-48b3-a391-58e0e611bfca/07aca0b7-dea4-48b3-a391-58e0e611bfca.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1088.874532] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2918e9c4-b8df-44f0-92e0-7ee883054d91 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.882267] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1088.882267] env[65522]: value = "task-5114761" [ 1088.882267] env[65522]: _type = "Task" [ 1088.882267] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.891689] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114761, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.043596] env[65522]: INFO nova.virt.block_device [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Booting with volume 3c40ed31-83a3-4a51-ace6-fae3fc2c638c at /dev/sdb [ 1089.085538] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-57fb14c1-a9cb-4bc2-a82a-a905a8547ba5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.096719] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af735086-5eda-4faf-9529-3e3429425bf8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.136861] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b666a5e-c494-4051-9260-6ef053363b8b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.147088] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8279911e-1264-4c3a-9efc-5c0104e5e40e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.170400] env[65522]: DEBUG nova.compute.utils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1089.190453] env[65522]: DEBUG nova.compute.manager [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1089.191626] env[65522]: DEBUG nova.network.neutron [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1089.191626] env[65522]: WARNING neutronclient.v2_0.client [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1089.192305] env[65522]: WARNING neutronclient.v2_0.client [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1089.194052] env[65522]: WARNING openstack [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1089.194052] env[65522]: WARNING openstack [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1089.206037] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7060f74f-cbb6-431f-950d-bc6c164dc85d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.215897] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0842439e-de58-4fda-af6b-046e9e649e4d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.234507] env[65522]: DEBUG nova.virt.block_device [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Updating existing volume attachment record: 30891edb-c0ec-414f-a65c-ba77772dfef1 {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1089.250282] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114760, 'name': CreateSnapshot_Task} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.250947] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1089.251821] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f2be18-9513-4249-9ea7-b6c167eaf820 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.268602] env[65522]: DEBUG nova.policy [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f51006be1b2e401a8dd5504e0e481810', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3ea0fab7de6c47eeaf3bdde413ae3901', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1089.303948] env[65522]: DEBUG oslo_vmware.api [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.361320] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1089.404320] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114761, 'name': Rename_Task, 'duration_secs': 0.251449} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.404826] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1089.405295] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9ce9923-9149-49ed-b1a8-0d3c7dd78590 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.420652] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1089.420652] env[65522]: value = "task-5114762" [ 1089.420652] env[65522]: _type = "Task" [ 1089.420652] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.438666] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114762, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.456614] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132c0a29-4afd-4a1f-af2a-5222f8d0f061 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.464099] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09602fd4-6816-417d-b7c7-0b922c9a6a33 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.499194] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e0d3de-af17-4493-b649-3bb34fd84df4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.510400] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95428208-30f4-4d90-a463-736b3fd3ee49 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.526408] env[65522]: DEBUG nova.compute.provider_tree [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1089.638358] env[65522]: DEBUG nova.network.neutron [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Successfully created port: 6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1089.673537] env[65522]: DEBUG nova.compute.manager [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1089.778995] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1089.782021] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9f37bfe8-d72e-4472-8240-6265e437a242 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.791755] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1089.791755] env[65522]: value = "task-5114763" [ 1089.791755] env[65522]: _type = "Task" [ 1089.791755] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.810425] env[65522]: DEBUG oslo_vmware.api [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.811349] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114763, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.932514] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114762, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.030617] env[65522]: DEBUG nova.scheduler.client.report [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1090.310931] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114763, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.316109] env[65522]: DEBUG oslo_vmware.api [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.433564] env[65522]: DEBUG oslo_vmware.api [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114762, 'name': PowerOnVM_Task, 'duration_secs': 0.789351} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.434035] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1090.434336] env[65522]: INFO nova.compute.manager [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Took 8.33 seconds to spawn the instance on the hypervisor. [ 1090.434555] env[65522]: DEBUG nova.compute.manager [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1090.435572] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7394bc-5bc6-414d-b8f7-201673fc52e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.537168] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.870s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1090.539552] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.178s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1090.541862] env[65522]: INFO nova.compute.claims [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1090.561954] env[65522]: INFO nova.scheduler.client.report [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleted allocations for instance 8859b051-8f75-4aad-b789-42662019d4c5 [ 1090.686099] env[65522]: DEBUG nova.compute.manager [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1090.713043] env[65522]: DEBUG nova.virt.hardware [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1090.713326] env[65522]: DEBUG nova.virt.hardware [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1090.713484] env[65522]: DEBUG nova.virt.hardware [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1090.713668] env[65522]: DEBUG nova.virt.hardware [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1090.713823] env[65522]: DEBUG nova.virt.hardware [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1090.713980] env[65522]: DEBUG nova.virt.hardware [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1090.714268] env[65522]: DEBUG nova.virt.hardware [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1090.714435] env[65522]: DEBUG nova.virt.hardware [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1090.714601] env[65522]: DEBUG nova.virt.hardware [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1090.714760] env[65522]: DEBUG nova.virt.hardware [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1090.714962] env[65522]: DEBUG nova.virt.hardware [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1090.715868] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b892cf2f-9225-49d3-9b99-5ba8c86b23f8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.725852] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a73461-cb70-4c83-a6fe-b444ea24a4a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.162334] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114763, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.162334] env[65522]: DEBUG oslo_vmware.api [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.173927] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1f1c65c9-af8c-44a8-ae76-28e407442737 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "8859b051-8f75-4aad-b789-42662019d4c5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.394s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1091.174825] env[65522]: INFO nova.compute.manager [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Took 16.63 seconds to build instance. [ 1091.308229] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114763, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.314991] env[65522]: DEBUG oslo_vmware.api [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.366111] env[65522]: DEBUG nova.compute.manager [req-24fff7f5-62f4-4d20-855a-7a97e077f702 req-fa3e13b9-5570-408e-9a26-b5e77987d547 service nova] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Received event network-vif-plugged-6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1091.366395] env[65522]: DEBUG oslo_concurrency.lockutils [req-24fff7f5-62f4-4d20-855a-7a97e077f702 req-fa3e13b9-5570-408e-9a26-b5e77987d547 service nova] Acquiring lock "45b4a56e-71b8-42ec-b925-a150e667b223-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1091.366630] env[65522]: DEBUG oslo_concurrency.lockutils [req-24fff7f5-62f4-4d20-855a-7a97e077f702 req-fa3e13b9-5570-408e-9a26-b5e77987d547 service nova] Lock "45b4a56e-71b8-42ec-b925-a150e667b223-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1091.366810] env[65522]: DEBUG oslo_concurrency.lockutils [req-24fff7f5-62f4-4d20-855a-7a97e077f702 req-fa3e13b9-5570-408e-9a26-b5e77987d547 service nova] Lock "45b4a56e-71b8-42ec-b925-a150e667b223-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1091.366978] env[65522]: DEBUG nova.compute.manager [req-24fff7f5-62f4-4d20-855a-7a97e077f702 req-fa3e13b9-5570-408e-9a26-b5e77987d547 service nova] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] No waiting events found dispatching network-vif-plugged-6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1091.367173] env[65522]: WARNING nova.compute.manager [req-24fff7f5-62f4-4d20-855a-7a97e077f702 req-fa3e13b9-5570-408e-9a26-b5e77987d547 service nova] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Received unexpected event network-vif-plugged-6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43 for instance with vm_state building and task_state spawning. [ 1091.495709] env[65522]: DEBUG nova.network.neutron [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Successfully updated port: 6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1091.678444] env[65522]: DEBUG oslo_concurrency.lockutils [None req-05792987-c51c-4198-bd0e-32c78a32d150 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "07aca0b7-dea4-48b3-a391-58e0e611bfca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.144s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1091.709114] env[65522]: DEBUG nova.virt.hardware [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1091.709422] env[65522]: DEBUG nova.virt.hardware [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1091.709592] env[65522]: DEBUG nova.virt.hardware [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1091.709701] env[65522]: DEBUG nova.virt.hardware [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1091.709843] env[65522]: DEBUG nova.virt.hardware [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1091.710186] env[65522]: DEBUG nova.virt.hardware [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1091.710434] env[65522]: DEBUG nova.virt.hardware [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1091.710602] env[65522]: DEBUG nova.virt.hardware [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1091.710871] env[65522]: DEBUG nova.virt.hardware [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1091.710969] env[65522]: DEBUG nova.virt.hardware [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1091.711109] env[65522]: DEBUG nova.virt.hardware [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1091.712484] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bacdf03-5e16-486a-ad61-8f43c708c389 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.723719] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c239891c-4839-4225-bec1-60ea3a71e284 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.742066] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:b9:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56136ef6-99d7-4562-9a9f-d66fec951c5c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '18edb16a-a08b-4e97-9b0a-a138798ed559', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1091.749700] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1091.752623] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1091.753125] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6d6fb88a-8b62-4d7c-a364-eb19f15f6258 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.778294] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1091.778294] env[65522]: value = "task-5114764" [ 1091.778294] env[65522]: _type = "Task" [ 1091.778294] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.788126] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114764, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.808033] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114763, 'name': CloneVM_Task, 'duration_secs': 1.767027} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.808333] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Created linked-clone VM from snapshot [ 1091.809563] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c317faad-71ec-475e-84f5-f835fef47934 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.817918] env[65522]: DEBUG oslo_vmware.api [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.823631] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Uploading image 3e1ae64d-4894-4ac1-8355-8aa9724ceee2 {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1091.849674] env[65522]: DEBUG oslo_vmware.rw_handles [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1091.849674] env[65522]: value = "vm-994928" [ 1091.849674] env[65522]: _type = "VirtualMachine" [ 1091.849674] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1091.849674] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-192298ee-6886-4ae1-a803-9923042d37d9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.861163] env[65522]: DEBUG oslo_vmware.rw_handles [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lease: (returnval){ [ 1091.861163] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5260f6a8-7fa7-06d3-678a-989584ab28cd" [ 1091.861163] env[65522]: _type = "HttpNfcLease" [ 1091.861163] env[65522]: } obtained for exporting VM: (result){ [ 1091.861163] env[65522]: value = "vm-994928" [ 1091.861163] env[65522]: _type = "VirtualMachine" [ 1091.861163] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1091.861570] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the lease: (returnval){ [ 1091.861570] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5260f6a8-7fa7-06d3-678a-989584ab28cd" [ 1091.861570] env[65522]: _type = "HttpNfcLease" [ 1091.861570] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1091.870711] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1091.870711] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5260f6a8-7fa7-06d3-678a-989584ab28cd" [ 1091.870711] env[65522]: _type = "HttpNfcLease" [ 1091.870711] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1091.916097] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dbd6baa-67ce-4ee0-b753-21ad6a740a15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.924650] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5927b2-d721-4238-b63a-b416e59fcf59 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.957966] env[65522]: DEBUG oslo_concurrency.lockutils [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "07aca0b7-dea4-48b3-a391-58e0e611bfca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1091.958246] env[65522]: DEBUG oslo_concurrency.lockutils [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "07aca0b7-dea4-48b3-a391-58e0e611bfca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1091.958474] env[65522]: DEBUG oslo_concurrency.lockutils [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "07aca0b7-dea4-48b3-a391-58e0e611bfca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1091.958671] env[65522]: DEBUG oslo_concurrency.lockutils [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "07aca0b7-dea4-48b3-a391-58e0e611bfca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1091.958841] env[65522]: DEBUG oslo_concurrency.lockutils [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "07aca0b7-dea4-48b3-a391-58e0e611bfca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1091.961317] env[65522]: INFO nova.compute.manager [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Terminating instance [ 1091.963347] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e09c7b0d-93ab-451f-a7c0-f0b0fc97a438 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.973583] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c819461e-07dd-468e-9e0b-df989c25972e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.992027] env[65522]: DEBUG nova.compute.provider_tree [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.997728] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "refresh_cache-45b4a56e-71b8-42ec-b925-a150e667b223" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.997910] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "refresh_cache-45b4a56e-71b8-42ec-b925-a150e667b223" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1091.998122] env[65522]: DEBUG nova.network.neutron [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1092.289177] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114764, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.312559] env[65522]: DEBUG oslo_vmware.api [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114752, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.371192] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1092.371192] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5260f6a8-7fa7-06d3-678a-989584ab28cd" [ 1092.371192] env[65522]: _type = "HttpNfcLease" [ 1092.371192] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1092.371663] env[65522]: DEBUG oslo_vmware.rw_handles [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1092.371663] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5260f6a8-7fa7-06d3-678a-989584ab28cd" [ 1092.371663] env[65522]: _type = "HttpNfcLease" [ 1092.371663] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1092.372307] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c259ad-319f-4ff5-831d-8ddca04b59f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.380518] env[65522]: DEBUG oslo_vmware.rw_handles [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52da528f-f9dc-bd51-e06f-168c1751d8cb/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1092.380709] env[65522]: DEBUG oslo_vmware.rw_handles [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52da528f-f9dc-bd51-e06f-168c1751d8cb/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1092.469217] env[65522]: DEBUG nova.compute.manager [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1092.469452] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1092.470607] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b443f91b-abf6-4d72-8c13-ba0242fc1c14 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.479848] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1092.480326] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80486c6c-6883-45a7-8cf6-c6fb2151ad7b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.486403] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ae8564b9-92f0-44b9-9854-c009ef3077ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.490626] env[65522]: DEBUG oslo_vmware.api [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1092.490626] env[65522]: value = "task-5114766" [ 1092.490626] env[65522]: _type = "Task" [ 1092.490626] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.497033] env[65522]: DEBUG nova.scheduler.client.report [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1092.502162] env[65522]: WARNING openstack [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1092.502709] env[65522]: WARNING openstack [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1092.520787] env[65522]: DEBUG oslo_vmware.api [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114766, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.557430] env[65522]: DEBUG nova.network.neutron [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1092.578831] env[65522]: WARNING openstack [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1092.579138] env[65522]: WARNING openstack [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1092.666329] env[65522]: WARNING neutronclient.v2_0.client [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1092.666989] env[65522]: WARNING openstack [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1092.667351] env[65522]: WARNING openstack [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1092.757837] env[65522]: DEBUG nova.network.neutron [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Updating instance_info_cache with network_info: [{"id": "6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43", "address": "fa:16:3e:b0:d8:1d", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fd7e155-ae", "ovs_interfaceid": "6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1092.790881] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114764, 'name': CreateVM_Task, 'duration_secs': 0.752654} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.791189] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1092.792072] env[65522]: WARNING neutronclient.v2_0.client [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 1092.792355] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.793187] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1092.793187] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1092.793187] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4184ead-a4ce-4e87-b287-f8b8bc071746 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.798713] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1092.798713] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b805bc-6ebf-a28a-2660-e5345a103f4d" [ 1092.798713] env[65522]: _type = "Task" [ 1092.798713] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.811761] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b805bc-6ebf-a28a-2660-e5345a103f4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.819457] env[65522]: DEBUG oslo_vmware.api [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114752, 'name': ReconfigVM_Task, 'duration_secs': 5.870664} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.819830] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1092.820117] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Reconfigured VM to detach interface {{(pid=65522) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1092.820581] env[65522]: WARNING neutronclient.v2_0.client [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1092.820902] env[65522]: WARNING neutronclient.v2_0.client [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1092.821504] env[65522]: WARNING openstack [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1092.821853] env[65522]: WARNING openstack [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1092.857435] env[65522]: WARNING neutronclient.v2_0.client [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.002645] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1093.003299] env[65522]: DEBUG nova.compute.manager [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1093.006260] env[65522]: DEBUG oslo_vmware.api [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114766, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.260942] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "refresh_cache-45b4a56e-71b8-42ec-b925-a150e667b223" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1093.261461] env[65522]: DEBUG nova.compute.manager [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Instance network_info: |[{"id": "6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43", "address": "fa:16:3e:b0:d8:1d", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fd7e155-ae", "ovs_interfaceid": "6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1093.261985] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:d8:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5e1049e8-c06b-4c93-a9e1-2cbb530f3f95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1093.269844] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1093.270254] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1093.270554] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-83fac208-a0bf-493a-b4b0-66c83e1b1f7f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.293695] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1093.293695] env[65522]: value = "task-5114767" [ 1093.293695] env[65522]: _type = "Task" [ 1093.293695] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.306943] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114767, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.315007] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b805bc-6ebf-a28a-2660-e5345a103f4d, 'name': SearchDatastore_Task, 'duration_secs': 0.059193} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.315381] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1093.315636] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1093.315921] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.316096] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1093.316296] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1093.316616] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d2e5221a-4773-4f56-8825-a61304cdb415 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.328890] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1093.331761] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1093.331761] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45b37066-060d-420a-90ed-9fef4f2d7107 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.337409] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1093.337409] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d3854a-7d5d-4019-c807-6ed0953bc001" [ 1093.337409] env[65522]: _type = "Task" [ 1093.337409] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.348741] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d3854a-7d5d-4019-c807-6ed0953bc001, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.429705] env[65522]: DEBUG nova.compute.manager [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Received event network-changed-6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1093.430153] env[65522]: DEBUG nova.compute.manager [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Refreshing instance network info cache due to event network-changed-6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1093.430501] env[65522]: DEBUG oslo_concurrency.lockutils [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] Acquiring lock "refresh_cache-45b4a56e-71b8-42ec-b925-a150e667b223" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.430661] env[65522]: DEBUG oslo_concurrency.lockutils [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] Acquired lock "refresh_cache-45b4a56e-71b8-42ec-b925-a150e667b223" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1093.430867] env[65522]: DEBUG nova.network.neutron [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Refreshing network info cache for port 6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1093.503081] env[65522]: DEBUG oslo_vmware.api [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114766, 'name': PowerOffVM_Task, 'duration_secs': 0.743008} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.503451] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1093.503678] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1093.504406] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e70a75b8-2c46-409b-a423-e85f103524be {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.509195] env[65522]: DEBUG nova.compute.utils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1093.511271] env[65522]: DEBUG nova.compute.manager [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1093.511364] env[65522]: DEBUG nova.network.neutron [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1093.511858] env[65522]: WARNING neutronclient.v2_0.client [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.512422] env[65522]: WARNING neutronclient.v2_0.client [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.513137] env[65522]: WARNING openstack [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.513545] env[65522]: WARNING openstack [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1093.573939] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1093.574787] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1093.574787] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleting the datastore file [datastore2] 07aca0b7-dea4-48b3-a391-58e0e611bfca {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1093.575136] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2db26de4-e337-4aae-9119-eafe29eb2db6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.583814] env[65522]: DEBUG oslo_vmware.api [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1093.583814] env[65522]: value = "task-5114769" [ 1093.583814] env[65522]: _type = "Task" [ 1093.583814] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.594844] env[65522]: DEBUG nova.policy [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18ceaff46b314c9f9da68a8d0c6f5f2b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fa11b46d9fe144f391233e6eb9c819d7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1093.602085] env[65522]: DEBUG oslo_vmware.api [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114769, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.805381] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114767, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.848414] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d3854a-7d5d-4019-c807-6ed0953bc001, 'name': SearchDatastore_Task, 'duration_secs': 0.015411} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.849294] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78dd49f9-f770-46ee-8f65-c7dbc42e406a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.856027] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1093.856027] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52be94fd-2822-b2c1-33bd-8e5c787dd588" [ 1093.856027] env[65522]: _type = "Task" [ 1093.856027] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.865470] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52be94fd-2822-b2c1-33bd-8e5c787dd588, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.907174] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1093.907492] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1093.907786] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1093.907928] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1093.908068] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1093.911202] env[65522]: INFO nova.compute.manager [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Terminating instance [ 1093.933708] env[65522]: DEBUG nova.network.neutron [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Successfully created port: a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1093.936476] env[65522]: WARNING neutronclient.v2_0.client [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.937169] env[65522]: WARNING openstack [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.937536] env[65522]: WARNING openstack [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.022644] env[65522]: DEBUG nova.compute.manager [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1094.049197] env[65522]: WARNING openstack [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.049711] env[65522]: WARNING openstack [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.096268] env[65522]: DEBUG oslo_vmware.api [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114769, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168078} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.096481] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1094.096481] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1094.096680] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1094.096814] env[65522]: INFO nova.compute.manager [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Took 1.63 seconds to destroy the instance on the hypervisor. [ 1094.097165] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1094.097280] env[65522]: DEBUG nova.compute.manager [-] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1094.097371] env[65522]: DEBUG nova.network.neutron [-] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1094.097612] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.098153] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.098465] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.116128] env[65522]: WARNING neutronclient.v2_0.client [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.116793] env[65522]: WARNING openstack [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.117172] env[65522]: WARNING openstack [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.146743] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.185974] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.186193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquired lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1094.186428] env[65522]: DEBUG nova.network.neutron [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1094.307725] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114767, 'name': CreateVM_Task, 'duration_secs': 0.81867} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.308068] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1094.308708] env[65522]: WARNING neutronclient.v2_0.client [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.309110] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.309318] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1094.309748] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1094.310112] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7f3fb07-75f0-45c0-b8eb-b6c6d8e816c3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.316175] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1094.316175] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525a97db-e614-f774-36b1-e4b45b34e935" [ 1094.316175] env[65522]: _type = "Task" [ 1094.316175] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.327026] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525a97db-e614-f774-36b1-e4b45b34e935, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.367927] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52be94fd-2822-b2c1-33bd-8e5c787dd588, 'name': SearchDatastore_Task, 'duration_secs': 0.016978} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.368338] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1094.368647] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 99353848-2f0f-4388-9fcd-91e799342386/99353848-2f0f-4388-9fcd-91e799342386.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1094.369020] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8d0f429e-451e-42e2-baff-373c49ba3d29 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.378246] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1094.378246] env[65522]: value = "task-5114770" [ 1094.378246] env[65522]: _type = "Task" [ 1094.378246] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.387621] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114770, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.420092] env[65522]: DEBUG nova.compute.manager [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1094.420285] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1094.421325] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e5df3c-69e8-4209-b18c-cd0359ca6fd4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.431060] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1094.431424] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28dcfdff-338f-4cb3-bd91-200b90b513bc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.439565] env[65522]: DEBUG oslo_vmware.api [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1094.439565] env[65522]: value = "task-5114771" [ 1094.439565] env[65522]: _type = "Task" [ 1094.439565] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.451360] env[65522]: DEBUG oslo_vmware.api [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114771, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.535755] env[65522]: DEBUG nova.compute.manager [req-0085713d-e1b9-43de-ba0b-155144993f15 req-3ba1e99c-68d9-4b69-ac99-184269a6552a service nova] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Received event network-vif-deleted-227f5a20-0177-480c-8cec-7485e6794c6d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1094.536065] env[65522]: INFO nova.compute.manager [req-0085713d-e1b9-43de-ba0b-155144993f15 req-3ba1e99c-68d9-4b69-ac99-184269a6552a service nova] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Neutron deleted interface 227f5a20-0177-480c-8cec-7485e6794c6d; detaching it from the instance and deleting it from the info cache [ 1094.536261] env[65522]: DEBUG nova.network.neutron [req-0085713d-e1b9-43de-ba0b-155144993f15 req-3ba1e99c-68d9-4b69-ac99-184269a6552a service nova] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1094.564941] env[65522]: DEBUG nova.network.neutron [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Updated VIF entry in instance network info cache for port 6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1094.565395] env[65522]: DEBUG nova.network.neutron [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Updating instance_info_cache with network_info: [{"id": "6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43", "address": "fa:16:3e:b0:d8:1d", "network": {"id": "70373599-786f-4e91-b949-ca9f2d2d8df1", "bridge": "br-int", "label": "tempest-ImagesTestJSON-643607710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3ea0fab7de6c47eeaf3bdde413ae3901", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fd7e155-ae", "ovs_interfaceid": "6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1094.689496] env[65522]: WARNING neutronclient.v2_0.client [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1094.690217] env[65522]: WARNING openstack [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.690580] env[65522]: WARNING openstack [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.834654] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525a97db-e614-f774-36b1-e4b45b34e935, 'name': SearchDatastore_Task, 'duration_secs': 0.024343} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.834654] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1094.835083] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1094.835193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.835326] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1094.835498] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1094.835809] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8a456be-ddc5-4774-aca9-f0683f5449c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.853168] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1094.853397] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1094.854262] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a625e44c-5582-46b3-97dd-3a225be0d04f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.863466] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1094.863466] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5206ad36-3aba-d466-071c-78e8e5359da8" [ 1094.863466] env[65522]: _type = "Task" [ 1094.863466] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.875062] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5206ad36-3aba-d466-071c-78e8e5359da8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.892575] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114770, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.931720] env[65522]: WARNING openstack [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1094.932052] env[65522]: WARNING openstack [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1094.953067] env[65522]: DEBUG oslo_vmware.api [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114771, 'name': PowerOffVM_Task, 'duration_secs': 0.250109} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.953400] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1094.953682] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1094.954109] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56bbd1b9-7938-4082-a011-2e7f165e6b4a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.004522] env[65522]: DEBUG nova.network.neutron [-] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1095.034154] env[65522]: DEBUG nova.compute.manager [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1095.036603] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1095.036878] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1095.037114] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Deleting the datastore file [datastore1] 95f0772d-6d3e-4e37-a0d2-9ab9de63d124 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1095.038089] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c2ebda8e-b10a-441e-988f-355ae68831ce {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.041429] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fbf2d570-73cf-48dd-bb4a-5d55bc809a1d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.047288] env[65522]: DEBUG oslo_vmware.api [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1095.047288] env[65522]: value = "task-5114773" [ 1095.047288] env[65522]: _type = "Task" [ 1095.047288] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.061038] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b33ef18-247c-496d-bca8-59728611810d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.076884] env[65522]: DEBUG oslo_concurrency.lockutils [req-15ee9201-5cdf-42fc-acd6-2cd2b3edfbcf req-e84e6c28-cc27-4665-a872-017ff124bc37 service nova] Releasing lock "refresh_cache-45b4a56e-71b8-42ec-b925-a150e667b223" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1095.077495] env[65522]: DEBUG oslo_vmware.api [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114773, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.110309] env[65522]: DEBUG nova.compute.manager [req-0085713d-e1b9-43de-ba0b-155144993f15 req-3ba1e99c-68d9-4b69-ac99-184269a6552a service nova] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Detach interface failed, port_id=227f5a20-0177-480c-8cec-7485e6794c6d, reason: Instance 07aca0b7-dea4-48b3-a391-58e0e611bfca could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1095.129371] env[65522]: WARNING neutronclient.v2_0.client [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1095.130167] env[65522]: WARNING openstack [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1095.130643] env[65522]: WARNING openstack [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1095.249134] env[65522]: INFO nova.network.neutron [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Port b6097e79-9e37-4e69-b0f3-40ddbe12d95a from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1095.249518] env[65522]: DEBUG nova.network.neutron [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updating instance_info_cache with network_info: [{"id": "d443e903-a2fa-4c64-b185-81fe54af293a", "address": "fa:16:3e:81:ed:9d", "network": {"id": "1ea0ec37-ea40-4911-b6ad-54375253cc42", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-541823482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "068c2387de8c406194d9b1762c7292a7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd443e903-a2", "ovs_interfaceid": "d443e903-a2fa-4c64-b185-81fe54af293a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1095.376404] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5206ad36-3aba-d466-071c-78e8e5359da8, 'name': SearchDatastore_Task, 'duration_secs': 0.056658} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.377543] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38ffaba1-4160-4978-b066-aec9358de9b3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.393406] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114770, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.64063} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.393818] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1095.393818] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526852af-9f79-1ade-b801-e14550694489" [ 1095.393818] env[65522]: _type = "Task" [ 1095.393818] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.394138] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 99353848-2f0f-4388-9fcd-91e799342386/99353848-2f0f-4388-9fcd-91e799342386.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1095.394372] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1095.394753] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-be579c03-e6dd-4c23-b0f4-676dd366f231 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.407618] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526852af-9f79-1ade-b801-e14550694489, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.409132] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1095.409132] env[65522]: value = "task-5114774" [ 1095.409132] env[65522]: _type = "Task" [ 1095.409132] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.418903] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114774, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.470484] env[65522]: DEBUG nova.compute.manager [req-846be735-fee9-4a87-805b-1c677df22db0 req-98b31887-4cb2-4108-85ef-10ba9e2b6947 service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Received event network-vif-plugged-a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1095.470955] env[65522]: DEBUG oslo_concurrency.lockutils [req-846be735-fee9-4a87-805b-1c677df22db0 req-98b31887-4cb2-4108-85ef-10ba9e2b6947 service nova] Acquiring lock "cb62df50-2f86-469a-b2ee-fd3754d61c83-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1095.471430] env[65522]: DEBUG oslo_concurrency.lockutils [req-846be735-fee9-4a87-805b-1c677df22db0 req-98b31887-4cb2-4108-85ef-10ba9e2b6947 service nova] Lock "cb62df50-2f86-469a-b2ee-fd3754d61c83-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1095.471797] env[65522]: DEBUG oslo_concurrency.lockutils [req-846be735-fee9-4a87-805b-1c677df22db0 req-98b31887-4cb2-4108-85ef-10ba9e2b6947 service nova] Lock "cb62df50-2f86-469a-b2ee-fd3754d61c83-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1095.472182] env[65522]: DEBUG nova.compute.manager [req-846be735-fee9-4a87-805b-1c677df22db0 req-98b31887-4cb2-4108-85ef-10ba9e2b6947 service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] No waiting events found dispatching network-vif-plugged-a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1095.472499] env[65522]: WARNING nova.compute.manager [req-846be735-fee9-4a87-805b-1c677df22db0 req-98b31887-4cb2-4108-85ef-10ba9e2b6947 service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Received unexpected event network-vif-plugged-a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac for instance with vm_state building and task_state spawning. [ 1095.498718] env[65522]: DEBUG nova.network.neutron [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Successfully updated port: a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1095.506964] env[65522]: INFO nova.compute.manager [-] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Took 1.41 seconds to deallocate network for instance. [ 1095.557680] env[65522]: DEBUG oslo_vmware.api [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114773, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.753167] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Releasing lock "refresh_cache-95f0772d-6d3e-4e37-a0d2-9ab9de63d124" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1095.907919] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526852af-9f79-1ade-b801-e14550694489, 'name': SearchDatastore_Task, 'duration_secs': 0.232838} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.908547] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1095.908972] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 45b4a56e-71b8-42ec-b925-a150e667b223/45b4a56e-71b8-42ec-b925-a150e667b223.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1095.909432] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a535f604-0eae-4631-ba87-f75474c85030 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.923401] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114774, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068681} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.925213] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1095.925833] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1095.925833] env[65522]: value = "task-5114775" [ 1095.925833] env[65522]: _type = "Task" [ 1095.925833] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.927033] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f7f13b-f817-4491-8be4-8035a2a6f3ac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.948877] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114775, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.958586] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 99353848-2f0f-4388-9fcd-91e799342386/99353848-2f0f-4388-9fcd-91e799342386.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1095.958924] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23799110-1ea6-4a91-baf1-0f90aee0ee96 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.980030] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1095.980030] env[65522]: value = "task-5114776" [ 1095.980030] env[65522]: _type = "Task" [ 1095.980030] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.990940] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114776, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.002966] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.003280] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1096.003475] env[65522]: DEBUG nova.network.neutron [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1096.015693] env[65522]: DEBUG oslo_concurrency.lockutils [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.015967] env[65522]: DEBUG oslo_concurrency.lockutils [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1096.016296] env[65522]: DEBUG nova.objects.instance [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lazy-loading 'resources' on Instance uuid 07aca0b7-dea4-48b3-a391-58e0e611bfca {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.061220] env[65522]: DEBUG oslo_vmware.api [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114773, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.782366} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.061539] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1096.061736] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1096.062015] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1096.062522] env[65522]: INFO nova.compute.manager [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Took 1.64 seconds to destroy the instance on the hypervisor. [ 1096.062638] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1096.062960] env[65522]: DEBUG nova.compute.manager [-] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1096.063214] env[65522]: DEBUG nova.network.neutron [-] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1096.063396] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1096.063967] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.064263] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.115860] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1096.257833] env[65522]: DEBUG oslo_concurrency.lockutils [None req-3569f540-a3ac-4a3c-b6a8-12840f049027 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "interface-95f0772d-6d3e-4e37-a0d2-9ab9de63d124-b6097e79-9e37-4e69-b0f3-40ddbe12d95a" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.092s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1096.442968] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114775, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.493461] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114776, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.507575] env[65522]: WARNING openstack [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.508245] env[65522]: WARNING openstack [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.558055] env[65522]: DEBUG nova.network.neutron [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1096.581322] env[65522]: WARNING openstack [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.581774] env[65522]: WARNING openstack [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.659574] env[65522]: WARNING neutronclient.v2_0.client [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1096.660361] env[65522]: WARNING openstack [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.660732] env[65522]: WARNING openstack [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.769476] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1494aede-391d-45a1-bb9b-da7409699ba2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.777140] env[65522]: DEBUG nova.network.neutron [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance_info_cache with network_info: [{"id": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "address": "fa:16:3e:5b:f2:d8", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7fc03fa-06", "ovs_interfaceid": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1096.781518] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46396fc7-6cb0-4677-90c4-0d2920ea4c96 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.814753] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59235455-e102-44d1-a072-9a04a80a00f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.823961] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4be2a6d-a9b5-43c6-b604-7a52b745a791 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.839921] env[65522]: DEBUG nova.compute.provider_tree [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1096.909635] env[65522]: DEBUG nova.compute.manager [req-0493876a-f0d1-4241-9199-517d45ead9e5 req-3f671c75-8002-48d2-920b-0575dc4bbb2b service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Received event network-vif-deleted-d443e903-a2fa-4c64-b185-81fe54af293a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1096.909892] env[65522]: INFO nova.compute.manager [req-0493876a-f0d1-4241-9199-517d45ead9e5 req-3f671c75-8002-48d2-920b-0575dc4bbb2b service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Neutron deleted interface d443e903-a2fa-4c64-b185-81fe54af293a; detaching it from the instance and deleting it from the info cache [ 1096.910129] env[65522]: DEBUG nova.network.neutron [req-0493876a-f0d1-4241-9199-517d45ead9e5 req-3f671c75-8002-48d2-920b-0575dc4bbb2b service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1096.942025] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114775, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.645455} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.942316] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 45b4a56e-71b8-42ec-b925-a150e667b223/45b4a56e-71b8-42ec-b925-a150e667b223.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1096.942612] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1096.943473] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c02c1148-5dda-4874-81ca-36825c453e8a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.952631] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1096.952631] env[65522]: value = "task-5114777" [ 1096.952631] env[65522]: _type = "Task" [ 1096.952631] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.963483] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114777, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.993481] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114776, 'name': ReconfigVM_Task, 'duration_secs': 0.77912} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.994019] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 99353848-2f0f-4388-9fcd-91e799342386/99353848-2f0f-4388-9fcd-91e799342386.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1096.995068] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_options': None, 'encrypted': False, 'boot_index': 0, 'disk_bus': None, 'size': 0, 'encryption_format': None, 'guest_format': None, 'encryption_secret_uuid': None, 'device_name': '/dev/sda', 'device_type': 'disk', 'image_id': 'a889db67-7337-4e32-8e34-642f34402926'}], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': False, 'boot_index': None, 'mount_device': '/dev/sdb', 'disk_bus': None, 'guest_format': None, 'device_type': None, 'attachment_id': '30891edb-c0ec-414f-a65c-ba77772dfef1', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994920', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'name': 'volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '99353848-2f0f-4388-9fcd-91e799342386', 'attached_at': '', 'detached_at': '', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'serial': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c'}, 'volume_type': None}], 'swap': None} {{(pid=65522) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1096.995305] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1096.995527] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994920', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'name': 'volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '99353848-2f0f-4388-9fcd-91e799342386', 'attached_at': '', 'detached_at': '', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'serial': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1096.996559] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f99a18-940a-47fc-9d1f-f6cd6a7b89e3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.016875] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f11afe-e3b9-4b74-9566-d2ab6c460b64 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.046291] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c/volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1097.047095] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a228a16-437b-49ad-8128-a37c4a14b0f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.068593] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1097.068593] env[65522]: value = "task-5114778" [ 1097.068593] env[65522]: _type = "Task" [ 1097.068593] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.078149] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114778, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.286666] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1097.286866] env[65522]: DEBUG nova.compute.manager [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Instance network_info: |[{"id": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "address": "fa:16:3e:5b:f2:d8", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7fc03fa-06", "ovs_interfaceid": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1097.344010] env[65522]: DEBUG nova.scheduler.client.report [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1097.385279] env[65522]: DEBUG nova.network.neutron [None req-c4044d4c-73bb-4db1-86d4-e8afb4aae1ee None None] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1097.413171] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bbc879b5-bb28-4f0e-9f4b-0738800f6104 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.425269] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd288ee-4cbb-4786-b979-54789311f270 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.468852] env[65522]: DEBUG nova.compute.manager [req-0493876a-f0d1-4241-9199-517d45ead9e5 req-3f671c75-8002-48d2-920b-0575dc4bbb2b service nova] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Detach interface failed, port_id=d443e903-a2fa-4c64-b185-81fe54af293a, reason: Instance 95f0772d-6d3e-4e37-a0d2-9ab9de63d124 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1097.481014] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114777, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079166} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.481508] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1097.482476] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f06a8438-03f4-4ae9-a3a8-8c86c0e93428 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.515082] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 45b4a56e-71b8-42ec-b925-a150e667b223/45b4a56e-71b8-42ec-b925-a150e667b223.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1097.515615] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35c8b94b-266f-4984-a878-d7c136619174 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.542115] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1097.542115] env[65522]: value = "task-5114779" [ 1097.542115] env[65522]: _type = "Task" [ 1097.542115] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.554239] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114779, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.565292] env[65522]: DEBUG nova.compute.manager [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Received event network-changed-a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1097.565548] env[65522]: DEBUG nova.compute.manager [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Refreshing instance network info cache due to event network-changed-a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1097.566139] env[65522]: DEBUG oslo_concurrency.lockutils [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] Acquiring lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.566139] env[65522]: DEBUG oslo_concurrency.lockutils [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] Acquired lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1097.566294] env[65522]: DEBUG nova.network.neutron [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Refreshing network info cache for port a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1097.581928] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114778, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.851329] env[65522]: DEBUG oslo_concurrency.lockutils [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.834s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1097.884897] env[65522]: INFO nova.scheduler.client.report [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleted allocations for instance 07aca0b7-dea4-48b3-a391-58e0e611bfca [ 1097.888845] env[65522]: INFO nova.compute.manager [None req-c4044d4c-73bb-4db1-86d4-e8afb4aae1ee None None] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Took 1.83 seconds to deallocate network for instance. [ 1098.054544] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114779, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.069557] env[65522]: WARNING neutronclient.v2_0.client [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.070526] env[65522]: WARNING openstack [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1098.071025] env[65522]: WARNING openstack [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1098.090519] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114778, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.176487] env[65522]: WARNING openstack [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1098.177022] env[65522]: WARNING openstack [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1098.247648] env[65522]: WARNING neutronclient.v2_0.client [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.249079] env[65522]: WARNING openstack [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1098.249079] env[65522]: WARNING openstack [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1098.345308] env[65522]: DEBUG nova.network.neutron [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updated VIF entry in instance network info cache for port a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1098.345779] env[65522]: DEBUG nova.network.neutron [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance_info_cache with network_info: [{"id": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "address": "fa:16:3e:5b:f2:d8", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7fc03fa-06", "ovs_interfaceid": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1098.392491] env[65522]: DEBUG oslo_concurrency.lockutils [None req-973ef294-8812-48c3-92b4-9e0ca9077267 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "07aca0b7-dea4-48b3-a391-58e0e611bfca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.434s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1098.397999] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.398283] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.398506] env[65522]: DEBUG nova.objects.instance [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'resources' on Instance uuid 95f0772d-6d3e-4e37-a0d2-9ab9de63d124 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1098.553797] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114779, 'name': ReconfigVM_Task, 'duration_secs': 0.770799} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.554287] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 45b4a56e-71b8-42ec-b925-a150e667b223/45b4a56e-71b8-42ec-b925-a150e667b223.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1098.554900] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d1fd575-073c-419a-a94e-c1f7039679b6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.563112] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1098.563112] env[65522]: value = "task-5114780" [ 1098.563112] env[65522]: _type = "Task" [ 1098.563112] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.572452] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114780, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.589544] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114778, 'name': ReconfigVM_Task, 'duration_secs': 1.306426} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.589894] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Reconfigured VM instance instance-00000058 to attach disk [datastore1] volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c/volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1098.595065] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07e07391-ba77-43c0-9ab9-1c8539cfbd87 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.612332] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1098.612332] env[65522]: value = "task-5114781" [ 1098.612332] env[65522]: _type = "Task" [ 1098.612332] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.622292] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114781, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.850358] env[65522]: DEBUG oslo_concurrency.lockutils [req-1ab471f2-7e63-46ce-9798-e52ff7a747c0 req-b2ef48a8-98b8-4b64-aec2-d841846641e6 service nova] Releasing lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1098.854066] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "210184da-48db-4926-ae00-7c0f29622810" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.854418] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "210184da-48db-4926-ae00-7c0f29622810" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.854666] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "210184da-48db-4926-ae00-7c0f29622810-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.854939] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "210184da-48db-4926-ae00-7c0f29622810-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.855162] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "210184da-48db-4926-ae00-7c0f29622810-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1098.858117] env[65522]: INFO nova.compute.manager [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Terminating instance [ 1099.077062] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114780, 'name': Rename_Task, 'duration_secs': 0.252036} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.077409] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1099.077727] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7836f02-7d3b-455f-bb03-8469cfce2b73 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.085240] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1099.085240] env[65522]: value = "task-5114782" [ 1099.085240] env[65522]: _type = "Task" [ 1099.085240] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.095963] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114782, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.097914] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef9c677-5e4a-4a05-ab54-30b05099e206 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.105986] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c85a2ea-0151-4d54-9231-e2bd7e87e36a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.143987] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a355a713-646e-4f06-be6a-5cd42a1c673d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.149807] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114781, 'name': ReconfigVM_Task, 'duration_secs': 0.289314} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.150710] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994920', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'name': 'volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '99353848-2f0f-4388-9fcd-91e799342386', 'attached_at': '', 'detached_at': '', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'serial': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1099.151230] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4135ed76-dedd-4ee2-abaa-76c0814441ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.158074] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c55999e-df07-4cd8-a606-922393a5b4bd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.163185] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1099.163185] env[65522]: value = "task-5114783" [ 1099.163185] env[65522]: _type = "Task" [ 1099.163185] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.177055] env[65522]: DEBUG nova.compute.provider_tree [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1099.186428] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114783, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.362646] env[65522]: DEBUG nova.compute.manager [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1099.362917] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1099.364041] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1173af61-a299-48fb-a4e2-e4136f8c62f8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.374915] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1099.375269] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c1bbe14-0480-4f85-bdbc-7ae5faaf8cd1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.383219] env[65522]: DEBUG oslo_vmware.api [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1099.383219] env[65522]: value = "task-5114784" [ 1099.383219] env[65522]: _type = "Task" [ 1099.383219] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.394948] env[65522]: DEBUG oslo_vmware.api [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114784, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.596756] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114782, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.676394] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114783, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.680680] env[65522]: DEBUG nova.scheduler.client.report [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1099.894551] env[65522]: DEBUG oslo_vmware.api [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114784, 'name': PowerOffVM_Task, 'duration_secs': 0.28993} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.894824] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1099.894988] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1099.895271] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e1e88255-4efa-43bc-b432-96eef843a376 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.965379] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1099.965697] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1099.965900] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleting the datastore file [datastore2] 210184da-48db-4926-ae00-7c0f29622810 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1099.966307] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af0470b7-c19a-4a13-8142-6522f5ded6b2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.974322] env[65522]: DEBUG oslo_vmware.api [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1099.974322] env[65522]: value = "task-5114786" [ 1099.974322] env[65522]: _type = "Task" [ 1099.974322] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.984597] env[65522]: DEBUG oslo_vmware.api [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114786, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.099302] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114782, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.175829] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114783, 'name': Rename_Task, 'duration_secs': 0.965925} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.176744] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1100.177012] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-659eb9e3-3bf7-4fe8-8a8f-8d137e18c376 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.183045] env[65522]: DEBUG nova.virt.hardware [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1100.183311] env[65522]: DEBUG nova.virt.hardware [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1100.183465] env[65522]: DEBUG nova.virt.hardware [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1100.183645] env[65522]: DEBUG nova.virt.hardware [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1100.183788] env[65522]: DEBUG nova.virt.hardware [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1100.183932] env[65522]: DEBUG nova.virt.hardware [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1100.184156] env[65522]: DEBUG nova.virt.hardware [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1100.184382] env[65522]: DEBUG nova.virt.hardware [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1100.184561] env[65522]: DEBUG nova.virt.hardware [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1100.184723] env[65522]: DEBUG nova.virt.hardware [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1100.184894] env[65522]: DEBUG nova.virt.hardware [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1100.186707] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.788s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1100.188996] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6040880-7fda-4dc4-90ed-4d17a0b5f4ce {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.193558] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1100.193558] env[65522]: value = "task-5114787" [ 1100.193558] env[65522]: _type = "Task" [ 1100.193558] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.201239] env[65522]: DEBUG oslo_vmware.rw_handles [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525e465f-a58c-6957-b27c-f4055735341a/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1100.203439] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ad3e31-0cd0-4efa-8342-233e96b25502 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.207878] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1964fdbe-a1ce-4ddc-a09a-e993bf56b497 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.216735] env[65522]: DEBUG oslo_vmware.rw_handles [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52da528f-f9dc-bd51-e06f-168c1751d8cb/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1100.221585] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8188eff4-484f-4942-a5e6-6a2e9076ef46 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.224215] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114787, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.225170] env[65522]: INFO nova.scheduler.client.report [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Deleted allocations for instance 95f0772d-6d3e-4e37-a0d2-9ab9de63d124 [ 1100.228438] env[65522]: DEBUG oslo_vmware.rw_handles [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525e465f-a58c-6957-b27c-f4055735341a/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1100.228438] env[65522]: ERROR oslo_vmware.rw_handles [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525e465f-a58c-6957-b27c-f4055735341a/disk-0.vmdk due to incomplete transfer. [ 1100.230762] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3de9a980-c8ed-4974-bf26-c53dbd69dbea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.242259] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:f2:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4307c18-b235-43cd-bcd5-e226012d8ee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1100.249764] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1100.250091] env[65522]: DEBUG oslo_vmware.rw_handles [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52da528f-f9dc-bd51-e06f-168c1751d8cb/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1100.250256] env[65522]: ERROR oslo_vmware.rw_handles [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52da528f-f9dc-bd51-e06f-168c1751d8cb/disk-0.vmdk due to incomplete transfer. [ 1100.250904] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1100.251136] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0394346f-21cf-4f40-87fc-5bdc8757fc6c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.252821] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6df9bab-25ea-4677-9754-3d9ead230a70 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.269379] env[65522]: DEBUG oslo_vmware.rw_handles [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525e465f-a58c-6957-b27c-f4055735341a/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1100.269573] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Uploaded image 9f4c00f3-d40c-4eeb-8605-5c426257ec5a to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1100.272388] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1100.273098] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-0beca9f5-4b22-4238-95e7-eff3c6645409 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.279030] env[65522]: DEBUG oslo_vmware.rw_handles [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52da528f-f9dc-bd51-e06f-168c1751d8cb/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1100.279030] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Uploaded image 3e1ae64d-4894-4ac1-8355-8aa9724ceee2 to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1100.279030] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1100.281658] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-330427de-ec19-4cef-b541-cc9ac9edeec4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.284037] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1100.284037] env[65522]: value = "task-5114788" [ 1100.284037] env[65522]: _type = "Task" [ 1100.284037] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.284037] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1100.284037] env[65522]: value = "task-5114789" [ 1100.284037] env[65522]: _type = "Task" [ 1100.284037] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.295766] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1100.295766] env[65522]: value = "task-5114790" [ 1100.295766] env[65522]: _type = "Task" [ 1100.295766] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.306350] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114788, 'name': CreateVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.307206] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114789, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.313238] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114790, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.486900] env[65522]: DEBUG oslo_vmware.api [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114786, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.378692} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.487218] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1100.487425] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1100.487624] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1100.487816] env[65522]: INFO nova.compute.manager [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 210184da-48db-4926-ae00-7c0f29622810] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1100.488116] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1100.488347] env[65522]: DEBUG nova.compute.manager [-] [instance: 210184da-48db-4926-ae00-7c0f29622810] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1100.488451] env[65522]: DEBUG nova.network.neutron [-] [instance: 210184da-48db-4926-ae00-7c0f29622810] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1100.488731] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1100.489374] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1100.489602] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1100.585995] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1100.598537] env[65522]: DEBUG oslo_vmware.api [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114782, 'name': PowerOnVM_Task, 'duration_secs': 1.491669} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.598910] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1100.600074] env[65522]: INFO nova.compute.manager [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Took 9.91 seconds to spawn the instance on the hypervisor. [ 1100.600074] env[65522]: DEBUG nova.compute.manager [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1100.600294] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d07214a-1d10-4fb7-9ae4-92ebb578adb3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.706700] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114787, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.736084] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4d39869e-0635-4221-824e-0e15edfac67d tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "95f0772d-6d3e-4e37-a0d2-9ab9de63d124" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.828s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1100.803292] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114789, 'name': Destroy_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.813612] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114788, 'name': CreateVM_Task, 'duration_secs': 0.442961} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.813612] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1100.813612] env[65522]: WARNING neutronclient.v2_0.client [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1100.813612] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.813612] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1100.813612] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1100.813612] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5428f24-0fce-4439-8824-147277d6bab2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.821103] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114790, 'name': Destroy_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.826080] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1100.826080] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c827b2-90e4-e80c-9d00-31ad814baae7" [ 1100.826080] env[65522]: _type = "Task" [ 1100.826080] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.837923] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c827b2-90e4-e80c-9d00-31ad814baae7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.028532] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1101.028847] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.029127] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1101.029357] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.029541] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.033673] env[65522]: INFO nova.compute.manager [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Terminating instance [ 1101.121428] env[65522]: INFO nova.compute.manager [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Took 18.82 seconds to build instance. [ 1101.206721] env[65522]: DEBUG oslo_vmware.api [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114787, 'name': PowerOnVM_Task, 'duration_secs': 0.957556} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.207077] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1101.207344] env[65522]: DEBUG nova.compute.manager [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1101.208174] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-004db5a1-4033-463d-9d1b-6e8bfea7ac7c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.270284] env[65522]: DEBUG nova.compute.manager [req-9abd7644-89db-4cd9-8e46-9409ed3cf25b req-0f02f6b9-2e3d-44f3-ae0d-d076d3761c8a service nova] [instance: 210184da-48db-4926-ae00-7c0f29622810] Received event network-vif-deleted-900d53cf-d483-46ab-8a04-c25218fc407c {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1101.270284] env[65522]: INFO nova.compute.manager [req-9abd7644-89db-4cd9-8e46-9409ed3cf25b req-0f02f6b9-2e3d-44f3-ae0d-d076d3761c8a service nova] [instance: 210184da-48db-4926-ae00-7c0f29622810] Neutron deleted interface 900d53cf-d483-46ab-8a04-c25218fc407c; detaching it from the instance and deleting it from the info cache [ 1101.270382] env[65522]: DEBUG nova.network.neutron [req-9abd7644-89db-4cd9-8e46-9409ed3cf25b req-0f02f6b9-2e3d-44f3-ae0d-d076d3761c8a service nova] [instance: 210184da-48db-4926-ae00-7c0f29622810] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1101.297257] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114789, 'name': Destroy_Task, 'duration_secs': 0.782047} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.297534] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Destroyed the VM [ 1101.298685] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1101.299316] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e138abd8-784c-44e0-bf3e-76f144a9db84 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.308443] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1101.308443] env[65522]: value = "task-5114791" [ 1101.308443] env[65522]: _type = "Task" [ 1101.308443] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.316332] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114790, 'name': Destroy_Task, 'duration_secs': 0.853767} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.316997] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Destroyed the VM [ 1101.317348] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1101.317622] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5675c42d-c431-4b36-a7ed-7baa48a6be81 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.322595] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114791, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.332225] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1101.332225] env[65522]: value = "task-5114792" [ 1101.332225] env[65522]: _type = "Task" [ 1101.332225] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.340241] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c827b2-90e4-e80c-9d00-31ad814baae7, 'name': SearchDatastore_Task, 'duration_secs': 0.049983} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.343760] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1101.344012] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1101.344256] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.344424] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1101.344612] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1101.344876] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114792, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.345111] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6e1086ef-c3b2-406c-9101-a7b44aaeebad {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.357371] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1101.357547] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1101.358332] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e5acaf7-266e-42fb-a5a4-9cbfe210b9b4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.364511] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1101.364511] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525ad8cc-8355-f0e0-5ba4-18fa4dfba2cf" [ 1101.364511] env[65522]: _type = "Task" [ 1101.364511] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.372982] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525ad8cc-8355-f0e0-5ba4-18fa4dfba2cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.545176] env[65522]: DEBUG nova.compute.manager [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1101.545176] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1101.546278] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8054f671-b2e7-4cbf-9d9f-0aa40446d69d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.555468] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1101.555766] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-875ed303-a317-4120-9148-b019db9e3cd4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.563619] env[65522]: DEBUG oslo_vmware.api [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1101.563619] env[65522]: value = "task-5114793" [ 1101.563619] env[65522]: _type = "Task" [ 1101.563619] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.575403] env[65522]: DEBUG oslo_vmware.api [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114793, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.624274] env[65522]: DEBUG oslo_concurrency.lockutils [None req-699ff1b3-185a-4ecc-9aac-1b5b07d16ff9 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "45b4a56e-71b8-42ec-b925-a150e667b223" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.329s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.730959] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1101.731161] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.731336] env[65522]: DEBUG nova.objects.instance [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1101.750819] env[65522]: DEBUG nova.network.neutron [-] [instance: 210184da-48db-4926-ae00-7c0f29622810] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1101.773596] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cbda6c14-4969-4bb7-90c9-45f9b1d3c0df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.788332] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3e53a1-6119-4421-ae22-1277366a6aa1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.827421] env[65522]: DEBUG nova.compute.manager [req-9abd7644-89db-4cd9-8e46-9409ed3cf25b req-0f02f6b9-2e3d-44f3-ae0d-d076d3761c8a service nova] [instance: 210184da-48db-4926-ae00-7c0f29622810] Detach interface failed, port_id=900d53cf-d483-46ab-8a04-c25218fc407c, reason: Instance 210184da-48db-4926-ae00-7c0f29622810 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1101.833518] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114791, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.842365] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114792, 'name': RemoveSnapshot_Task, 'duration_secs': 0.476925} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.842657] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1101.842955] env[65522]: DEBUG nova.compute.manager [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1101.843821] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d7483e-1cef-4752-abe5-5f4e97d595da {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.876287] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525ad8cc-8355-f0e0-5ba4-18fa4dfba2cf, 'name': SearchDatastore_Task, 'duration_secs': 0.014327} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.877115] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3579df03-751e-4ea0-b3da-f67bc0780cd9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.883420] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1101.883420] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5217fd36-f12f-65fc-1f8e-dec29cdd2171" [ 1101.883420] env[65522]: _type = "Task" [ 1101.883420] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.892638] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5217fd36-f12f-65fc-1f8e-dec29cdd2171, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.075258] env[65522]: DEBUG oslo_vmware.api [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114793, 'name': PowerOffVM_Task, 'duration_secs': 0.237878} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.076075] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1102.076251] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1102.076507] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4e01d50-f643-4ccc-9985-06f0c920f667 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.139218] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1102.139434] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1102.139609] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Deleting the datastore file [datastore1] 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1102.139905] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4a3ea4c-dcfb-4166-822c-d54001fc4af6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.147943] env[65522]: DEBUG oslo_vmware.api [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for the task: (returnval){ [ 1102.147943] env[65522]: value = "task-5114795" [ 1102.147943] env[65522]: _type = "Task" [ 1102.147943] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.158905] env[65522]: DEBUG oslo_vmware.api [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114795, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.253547] env[65522]: INFO nova.compute.manager [-] [instance: 210184da-48db-4926-ae00-7c0f29622810] Took 1.76 seconds to deallocate network for instance. [ 1102.320424] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114791, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.357554] env[65522]: INFO nova.compute.manager [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Shelve offloading [ 1102.395046] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5217fd36-f12f-65fc-1f8e-dec29cdd2171, 'name': SearchDatastore_Task, 'duration_secs': 0.027778} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.395339] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1102.395598] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] cb62df50-2f86-469a-b2ee-fd3754d61c83/cb62df50-2f86-469a-b2ee-fd3754d61c83.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1102.395866] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a80bc24-cab6-4f5a-80d5-fabfe78d8596 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.402935] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1102.402935] env[65522]: value = "task-5114796" [ 1102.402935] env[65522]: _type = "Task" [ 1102.402935] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.411675] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114796, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.659856] env[65522]: DEBUG oslo_vmware.api [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Task: {'id': task-5114795, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.301923} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.660182] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1102.660380] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1102.660560] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1102.660725] env[65522]: INFO nova.compute.manager [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1102.660968] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1102.661206] env[65522]: DEBUG nova.compute.manager [-] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1102.661329] env[65522]: DEBUG nova.network.neutron [-] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1102.661578] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1102.662173] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1102.662485] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1102.698544] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1102.741376] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7de4ec42-f1f5-4843-8dd6-b393a7a5bbed tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1102.761272] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1102.761541] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1102.761763] env[65522]: DEBUG nova.objects.instance [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lazy-loading 'resources' on Instance uuid 210184da-48db-4926-ae00-7c0f29622810 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.821464] env[65522]: DEBUG oslo_vmware.api [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114791, 'name': RemoveSnapshot_Task, 'duration_secs': 1.019515} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.821794] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1102.822062] env[65522]: INFO nova.compute.manager [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Took 19.28 seconds to snapshot the instance on the hypervisor. [ 1102.861409] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1102.861781] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8636f751-d2db-475f-a669-be3643d758a3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.870309] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1102.870309] env[65522]: value = "task-5114797" [ 1102.870309] env[65522]: _type = "Task" [ 1102.870309] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.880186] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114797, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.914289] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114796, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.938817] env[65522]: DEBUG nova.compute.manager [req-e21ffc8c-add2-4e10-b0e5-bf954589b421 req-c11cb99a-0681-4e47-bf29-c3e7b8106f09 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Received event network-vif-deleted-2b2de45d-4c08-4bb4-810a-2f26266cbc20 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1102.939107] env[65522]: INFO nova.compute.manager [req-e21ffc8c-add2-4e10-b0e5-bf954589b421 req-c11cb99a-0681-4e47-bf29-c3e7b8106f09 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Neutron deleted interface 2b2de45d-4c08-4bb4-810a-2f26266cbc20; detaching it from the instance and deleting it from the info cache [ 1102.939268] env[65522]: DEBUG nova.network.neutron [req-e21ffc8c-add2-4e10-b0e5-bf954589b421 req-c11cb99a-0681-4e47-bf29-c3e7b8106f09 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1103.072102] env[65522]: DEBUG nova.compute.manager [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1103.073613] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b10ac5a-4314-44d6-90bc-be9ff1c00cdd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.392638] env[65522]: DEBUG nova.compute.manager [None req-edc3b1f1-72a5-4641-b0d2-64c40d306a57 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Found 2 images (rotation: 2) {{(pid=65522) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 1103.393913] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1103.394182] env[65522]: DEBUG nova.compute.manager [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1103.395719] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-582b21f4-7f02-4d03-9a9a-0926f3101f98 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.405982] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "refresh_cache-821a5847-c255-4846-9041-2a0144f4539b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.406185] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "refresh_cache-821a5847-c255-4846-9041-2a0144f4539b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1103.406356] env[65522]: DEBUG nova.network.neutron [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1103.421802] env[65522]: DEBUG nova.network.neutron [-] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1103.422980] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114796, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.441906] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b644ef47-ac2a-47e5-b3bf-5cc4d2a1f3b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.457124] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42670629-8521-4ed6-a36c-0f17ee01f5bc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.483705] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1e7228-c0c9-4dae-bb53-4e32ba3c2e2a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.500675] env[65522]: DEBUG nova.compute.manager [req-e21ffc8c-add2-4e10-b0e5-bf954589b421 req-c11cb99a-0681-4e47-bf29-c3e7b8106f09 service nova] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Detach interface failed, port_id=2b2de45d-4c08-4bb4-810a-2f26266cbc20, reason: Instance 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1103.507879] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73afac50-5953-45f6-9030-bf1cfa25adc4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.541993] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229e651e-2f25-45d0-ac57-72cc42e9837d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.550068] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f3b0f62-011a-44db-9e64-4e4457dea82e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.565714] env[65522]: DEBUG nova.compute.provider_tree [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1103.585768] env[65522]: INFO nova.compute.manager [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] instance snapshotting [ 1103.588455] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b401e1f-1dd2-4976-9210-7e7dfa797dc6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.609260] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773f8de1-c2bd-4614-89ee-139812de7454 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.912810] env[65522]: WARNING neutronclient.v2_0.client [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1103.913935] env[65522]: WARNING openstack [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.913935] env[65522]: WARNING openstack [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.926648] env[65522]: INFO nova.compute.manager [-] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Took 1.27 seconds to deallocate network for instance. [ 1103.927174] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114796, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.020248} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.931513] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] cb62df50-2f86-469a-b2ee-fd3754d61c83/cb62df50-2f86-469a-b2ee-fd3754d61c83.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1103.932355] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1103.932355] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-248373c2-139c-4bc9-9c34-bce807fbec70 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.941333] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1103.941333] env[65522]: value = "task-5114798" [ 1103.941333] env[65522]: _type = "Task" [ 1103.941333] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.952791] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114798, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.032281] env[65522]: WARNING openstack [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.032635] env[65522]: WARNING openstack [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.069621] env[65522]: DEBUG nova.scheduler.client.report [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1104.104776] env[65522]: WARNING neutronclient.v2_0.client [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1104.105573] env[65522]: WARNING openstack [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.105943] env[65522]: WARNING openstack [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.121036] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1104.121283] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e749b090-df8d-43d5-b571-0656b85d014a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.130320] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1104.130320] env[65522]: value = "task-5114799" [ 1104.130320] env[65522]: _type = "Task" [ 1104.130320] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.140423] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114799, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.204063] env[65522]: DEBUG nova.network.neutron [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Updating instance_info_cache with network_info: [{"id": "0ef0d28e-583d-479c-b4f1-782fe7ad45f2", "address": "fa:16:3e:b0:ef:01", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ef0d28e-58", "ovs_interfaceid": "0ef0d28e-583d-479c-b4f1-782fe7ad45f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1104.434812] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1104.454615] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114798, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.197363} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.454907] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1104.455760] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e625402c-fc2f-499c-beb5-8f69dc1cd787 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.483096] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] cb62df50-2f86-469a-b2ee-fd3754d61c83/cb62df50-2f86-469a-b2ee-fd3754d61c83.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1104.483874] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-30a6211d-c498-4c40-a043-e84941b7e692 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.506232] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1104.506232] env[65522]: value = "task-5114800" [ 1104.506232] env[65522]: _type = "Task" [ 1104.506232] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.516811] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114800, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.575324] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.813s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1104.578362] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.144s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1104.578702] env[65522]: DEBUG nova.objects.instance [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lazy-loading 'resources' on Instance uuid 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1104.600662] env[65522]: INFO nova.scheduler.client.report [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleted allocations for instance 210184da-48db-4926-ae00-7c0f29622810 [ 1104.641922] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114799, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.683250] env[65522]: DEBUG nova.compute.manager [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1104.684253] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6feeac8b-ec78-4d77-bc67-c4a002dd7662 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.707243] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "refresh_cache-821a5847-c255-4846-9041-2a0144f4539b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1104.707428] env[65522]: WARNING neutronclient.v2_0.client [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1104.708014] env[65522]: WARNING openstack [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1104.708386] env[65522]: WARNING openstack [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1104.713872] env[65522]: WARNING neutronclient.v2_0.client [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1105.007072] env[65522]: DEBUG nova.compute.manager [req-f1c23d4c-05c5-4b73-a1f8-bb2bb19de8bd req-66a2a22a-afc8-440a-905d-7d341e03e83e service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Received event network-vif-unplugged-0ef0d28e-583d-479c-b4f1-782fe7ad45f2 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1105.007072] env[65522]: DEBUG oslo_concurrency.lockutils [req-f1c23d4c-05c5-4b73-a1f8-bb2bb19de8bd req-66a2a22a-afc8-440a-905d-7d341e03e83e service nova] Acquiring lock "821a5847-c255-4846-9041-2a0144f4539b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1105.007675] env[65522]: DEBUG oslo_concurrency.lockutils [req-f1c23d4c-05c5-4b73-a1f8-bb2bb19de8bd req-66a2a22a-afc8-440a-905d-7d341e03e83e service nova] Lock "821a5847-c255-4846-9041-2a0144f4539b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1105.007675] env[65522]: DEBUG oslo_concurrency.lockutils [req-f1c23d4c-05c5-4b73-a1f8-bb2bb19de8bd req-66a2a22a-afc8-440a-905d-7d341e03e83e service nova] Lock "821a5847-c255-4846-9041-2a0144f4539b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1105.007675] env[65522]: DEBUG nova.compute.manager [req-f1c23d4c-05c5-4b73-a1f8-bb2bb19de8bd req-66a2a22a-afc8-440a-905d-7d341e03e83e service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] No waiting events found dispatching network-vif-unplugged-0ef0d28e-583d-479c-b4f1-782fe7ad45f2 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1105.007675] env[65522]: WARNING nova.compute.manager [req-f1c23d4c-05c5-4b73-a1f8-bb2bb19de8bd req-66a2a22a-afc8-440a-905d-7d341e03e83e service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Received unexpected event network-vif-unplugged-0ef0d28e-583d-479c-b4f1-782fe7ad45f2 for instance with vm_state shelved and task_state shelving_offloading. [ 1105.020199] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114800, 'name': ReconfigVM_Task, 'duration_secs': 0.323241} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.020517] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Reconfigured VM instance instance-00000064 to attach disk [datastore2] cb62df50-2f86-469a-b2ee-fd3754d61c83/cb62df50-2f86-469a-b2ee-fd3754d61c83.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1105.021309] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b89876b1-5f90-4aa6-908a-f4e7a169c5e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.031972] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1105.031972] env[65522]: value = "task-5114801" [ 1105.031972] env[65522]: _type = "Task" [ 1105.031972] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.033166] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1105.035044] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8cff1f-6702-419b-94d8-351b2d339158 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.045310] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1105.049734] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a3848b24-dd34-4da8-bc25-e0f3c367c2c2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.050994] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114801, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.109656] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8f217e49-1810-46ba-ab60-4282932d631c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "210184da-48db-4926-ae00-7c0f29622810" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.255s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1105.119720] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1105.119989] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1105.120204] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleting the datastore file [datastore1] 821a5847-c255-4846-9041-2a0144f4539b {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1105.120506] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14be27f1-d6ed-4eaa-9499-dd6df1e90606 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.133750] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1105.133750] env[65522]: value = "task-5114803" [ 1105.133750] env[65522]: _type = "Task" [ 1105.133750] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.151277] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114803, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.151874] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114799, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.196613] env[65522]: INFO nova.compute.manager [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] instance snapshotting [ 1105.197294] env[65522]: DEBUG nova.objects.instance [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'flavor' on Instance uuid 9f342b89-bde2-4c35-ae42-cfe1e6973b74 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.250771] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac2d65be-e93d-4f60-912c-f2702e9b4956 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.258433] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c605b926-8905-492e-8b47-2d58097bd1c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.293113] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c45ef280-4eeb-43df-88a3-cbee511a6717 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.301639] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d69e801-0a1b-49f2-9841-ba5201d6277a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.316651] env[65522]: DEBUG nova.compute.provider_tree [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1105.544554] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114801, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.650813] env[65522]: DEBUG oslo_vmware.api [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114803, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.238114} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.652237] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114799, 'name': CreateSnapshot_Task, 'duration_secs': 1.426078} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.652237] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1105.652237] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1105.652237] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1105.653815] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1105.654799] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283d5492-c725-4f84-9cdb-1c8424cd97c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.677757] env[65522]: INFO nova.scheduler.client.report [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleted allocations for instance 821a5847-c255-4846-9041-2a0144f4539b [ 1105.705507] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da47b74-d4c2-47b2-8b02-5f6ec551fe59 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.740163] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e62ff0e-8879-45f7-be37-ee252376fdbd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.819969] env[65522]: DEBUG nova.scheduler.client.report [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1106.045296] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114801, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.129745] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "2b61302b-7ce1-4e8d-b981-8d4376d76c3c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1106.130119] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "2b61302b-7ce1-4e8d-b981-8d4376d76c3c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1106.175075] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1106.175669] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-982d0667-5982-4cce-af2d-74378145d9e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.184438] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1106.184438] env[65522]: value = "task-5114804" [ 1106.184438] env[65522]: _type = "Task" [ 1106.184438] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.188721] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1106.194151] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114804, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.252420] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1106.252859] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d6a8c3f2-5b81-491b-8f86-b510d9445291 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.261039] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1106.261039] env[65522]: value = "task-5114805" [ 1106.261039] env[65522]: _type = "Task" [ 1106.261039] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.270756] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114805, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.325613] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.747s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1106.328245] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.139s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1106.328494] env[65522]: DEBUG nova.objects.instance [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lazy-loading 'resources' on Instance uuid 821a5847-c255-4846-9041-2a0144f4539b {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1106.343210] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1106.343533] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1106.350670] env[65522]: INFO nova.scheduler.client.report [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Deleted allocations for instance 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3 [ 1106.548681] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114801, 'name': Rename_Task, 'duration_secs': 1.114861} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.549073] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1106.549387] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73b2f907-2d8d-4c9c-a117-12097255f8ea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.557390] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1106.557390] env[65522]: value = "task-5114806" [ 1106.557390] env[65522]: _type = "Task" [ 1106.557390] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.567171] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114806, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.633072] env[65522]: DEBUG nova.compute.manager [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1106.698493] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114804, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.772142] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114805, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.831802] env[65522]: DEBUG nova.objects.instance [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lazy-loading 'numa_topology' on Instance uuid 821a5847-c255-4846-9041-2a0144f4539b {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1106.847897] env[65522]: INFO nova.compute.manager [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Detaching volume a517fee4-4f9a-40e5-9c5c-0240aebef0cc [ 1106.860843] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cd53af57-0350-4ac0-b62a-e621298493a7 tempest-AttachInterfacesTestJSON-1046622067 tempest-AttachInterfacesTestJSON-1046622067-project-member] Lock "11ef190f-8ade-4705-b3b0-a8ff8b97bcd3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.832s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1106.889177] env[65522]: INFO nova.virt.block_device [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Attempting to driver detach volume a517fee4-4f9a-40e5-9c5c-0240aebef0cc from mountpoint /dev/sdb [ 1106.889479] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1106.889765] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994906', 'volume_id': 'a517fee4-4f9a-40e5-9c5c-0240aebef0cc', 'name': 'volume-a517fee4-4f9a-40e5-9c5c-0240aebef0cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5ce4a286-efd1-4bbc-a23b-931c6701cfe4', 'attached_at': '', 'detached_at': '', 'volume_id': 'a517fee4-4f9a-40e5-9c5c-0240aebef0cc', 'serial': 'a517fee4-4f9a-40e5-9c5c-0240aebef0cc'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1106.890792] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93dbaf9d-b263-4336-a927-ad19939024ac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.915161] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffdafc48-1baa-4582-941b-7612e9ca6582 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.922936] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef6018f-602e-4f0d-8291-2f3cf5b4cd87 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.948824] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1dc605-e575-4a67-9603-fe50d790efc1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.966833] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The volume has not been displaced from its original location: [datastore1] volume-a517fee4-4f9a-40e5-9c5c-0240aebef0cc/volume-a517fee4-4f9a-40e5-9c5c-0240aebef0cc.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1106.973965] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Reconfiguring VM instance instance-00000045 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1106.974398] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d564b320-da84-4f98-9a6f-4695101c88d6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.998426] env[65522]: DEBUG oslo_vmware.api [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1106.998426] env[65522]: value = "task-5114807" [ 1106.998426] env[65522]: _type = "Task" [ 1106.998426] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.010638] env[65522]: DEBUG oslo_vmware.api [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114807, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.041059] env[65522]: DEBUG nova.compute.manager [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Received event network-changed-0ef0d28e-583d-479c-b4f1-782fe7ad45f2 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1107.041327] env[65522]: DEBUG nova.compute.manager [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Refreshing instance network info cache due to event network-changed-0ef0d28e-583d-479c-b4f1-782fe7ad45f2. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1107.041684] env[65522]: DEBUG oslo_concurrency.lockutils [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] Acquiring lock "refresh_cache-821a5847-c255-4846-9041-2a0144f4539b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.041760] env[65522]: DEBUG oslo_concurrency.lockutils [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] Acquired lock "refresh_cache-821a5847-c255-4846-9041-2a0144f4539b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1107.041914] env[65522]: DEBUG nova.network.neutron [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Refreshing network info cache for port 0ef0d28e-583d-479c-b4f1-782fe7ad45f2 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1107.068235] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114806, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.155881] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1107.198315] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114804, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.247123] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c3a86cb-baf4-48dd-91f2-54560de66cf5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "821a5847-c255-4846-9041-2a0144f4539b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1107.274654] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114805, 'name': CreateSnapshot_Task, 'duration_secs': 0.941166} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.274986] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1107.275760] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39560f28-56f1-45c9-95f4-c2e7ab119a1a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.334281] env[65522]: DEBUG nova.objects.base [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Object Instance<821a5847-c255-4846-9041-2a0144f4539b> lazy-loaded attributes: resources,numa_topology {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1107.498407] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990e856f-403d-4eb2-a4f7-476a6bf96df9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.510812] env[65522]: DEBUG oslo_vmware.api [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114807, 'name': ReconfigVM_Task, 'duration_secs': 0.386438} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.512534] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Reconfigured VM instance instance-00000045 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1107.517245] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40831caa-6c70-4ba0-835d-d17fc20b1c98 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.527919] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019539cf-7a8a-48f2-90ec-1e1f13bdbe62 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.536395] env[65522]: DEBUG oslo_vmware.api [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1107.536395] env[65522]: value = "task-5114808" [ 1107.536395] env[65522]: _type = "Task" [ 1107.536395] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.563040] env[65522]: WARNING neutronclient.v2_0.client [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.563956] env[65522]: WARNING openstack [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.565127] env[65522]: WARNING openstack [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.582846] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e2d225-ddef-497c-86de-08bdff7fd540 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.593708] env[65522]: DEBUG oslo_vmware.api [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114808, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.599710] env[65522]: DEBUG oslo_vmware.api [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114806, 'name': PowerOnVM_Task, 'duration_secs': 0.526732} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.600279] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0d6e9e-af2b-4b83-b83f-43ae64a812f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.608046] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1107.608046] env[65522]: INFO nova.compute.manager [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Took 12.57 seconds to spawn the instance on the hypervisor. [ 1107.608046] env[65522]: DEBUG nova.compute.manager [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1107.610341] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda80cd4-5742-496e-9ceb-19cb5ac94e4f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.624354] env[65522]: DEBUG nova.compute.provider_tree [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1107.678898] env[65522]: WARNING openstack [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.679427] env[65522]: WARNING openstack [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.697249] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114804, 'name': CloneVM_Task, 'duration_secs': 1.413605} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.697384] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Created linked-clone VM from snapshot [ 1107.698199] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f976b3-526b-4bad-be52-87f51c3b361c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.708207] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Uploading image fd2d6661-5057-4e8e-82c2-1523528553b2 {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1107.722776] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1107.723118] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a1a42836-4958-414b-84d2-56a2c4b533a1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.731805] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1107.731805] env[65522]: value = "task-5114809" [ 1107.731805] env[65522]: _type = "Task" [ 1107.731805] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.742205] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114809, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.745537] env[65522]: WARNING neutronclient.v2_0.client [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.746267] env[65522]: WARNING openstack [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.746678] env[65522]: WARNING openstack [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.796362] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1107.796362] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0e27a43d-142d-4a62-9603-3275e21040d8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.805885] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1107.805885] env[65522]: value = "task-5114810" [ 1107.805885] env[65522]: _type = "Task" [ 1107.805885] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.821102] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114810, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.876437] env[65522]: DEBUG nova.network.neutron [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Updated VIF entry in instance network info cache for port 0ef0d28e-583d-479c-b4f1-782fe7ad45f2. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1107.877112] env[65522]: DEBUG nova.network.neutron [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Updating instance_info_cache with network_info: [{"id": "0ef0d28e-583d-479c-b4f1-782fe7ad45f2", "address": "fa:16:3e:b0:ef:01", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": null, "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap0ef0d28e-58", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1108.073447] env[65522]: DEBUG oslo_vmware.api [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114808, 'name': ReconfigVM_Task, 'duration_secs': 0.178883} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.073898] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994906', 'volume_id': 'a517fee4-4f9a-40e5-9c5c-0240aebef0cc', 'name': 'volume-a517fee4-4f9a-40e5-9c5c-0240aebef0cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5ce4a286-efd1-4bbc-a23b-931c6701cfe4', 'attached_at': '', 'detached_at': '', 'volume_id': 'a517fee4-4f9a-40e5-9c5c-0240aebef0cc', 'serial': 'a517fee4-4f9a-40e5-9c5c-0240aebef0cc'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1108.133085] env[65522]: DEBUG nova.scheduler.client.report [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1108.148074] env[65522]: INFO nova.compute.manager [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Took 18.81 seconds to build instance. [ 1108.244435] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114809, 'name': Destroy_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.316199] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114810, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.380394] env[65522]: DEBUG oslo_concurrency.lockutils [req-7b29c792-0b23-4f19-9071-2dd3c1914e01 req-ec3e0130-fda7-4b8a-b58a-874bbe7ad0d4 service nova] Releasing lock "refresh_cache-821a5847-c255-4846-9041-2a0144f4539b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1108.620722] env[65522]: DEBUG nova.objects.instance [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lazy-loading 'flavor' on Instance uuid 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1108.644088] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.316s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1108.649608] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.491s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1108.649608] env[65522]: INFO nova.compute.claims [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1108.650907] env[65522]: DEBUG oslo_concurrency.lockutils [None req-fe8b61b1-6d54-464e-8d94-3a31ed448e58 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "cb62df50-2f86-469a-b2ee-fd3754d61c83" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.322s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1108.743731] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114809, 'name': Destroy_Task, 'duration_secs': 0.565782} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.744015] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Destroyed the VM [ 1108.744254] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1108.744554] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3d2b64e8-b88b-4e3b-b18a-32a37f4f0709 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.752139] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1108.752139] env[65522]: value = "task-5114811" [ 1108.752139] env[65522]: _type = "Task" [ 1108.752139] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.764049] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114811, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.817925] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114810, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.076667] env[65522]: DEBUG nova.compute.manager [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Received event network-changed-a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1109.076667] env[65522]: DEBUG nova.compute.manager [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Refreshing instance network info cache due to event network-changed-a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1109.076667] env[65522]: DEBUG oslo_concurrency.lockutils [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] Acquiring lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.077082] env[65522]: DEBUG oslo_concurrency.lockutils [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] Acquired lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1109.077082] env[65522]: DEBUG nova.network.neutron [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Refreshing network info cache for port a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1109.169010] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bb6db97d-f044-4e1f-acd8-81eb2083497f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "821a5847-c255-4846-9041-2a0144f4539b" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.049s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.171085] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c3a86cb-baf4-48dd-91f2-54560de66cf5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "821a5847-c255-4846-9041-2a0144f4539b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 1.923s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1109.171085] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c3a86cb-baf4-48dd-91f2-54560de66cf5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "821a5847-c255-4846-9041-2a0144f4539b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1109.171085] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c3a86cb-baf4-48dd-91f2-54560de66cf5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "821a5847-c255-4846-9041-2a0144f4539b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1109.171085] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c3a86cb-baf4-48dd-91f2-54560de66cf5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "821a5847-c255-4846-9041-2a0144f4539b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.174090] env[65522]: INFO nova.compute.manager [None req-6c3a86cb-baf4-48dd-91f2-54560de66cf5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Terminating instance [ 1109.267297] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114811, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.326972] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114810, 'name': CloneVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.552666] env[65522]: DEBUG oslo_concurrency.lockutils [None req-eab79372-f48d-4fb3-839f-6b68f0044ce6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1109.581782] env[65522]: WARNING neutronclient.v2_0.client [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1109.582785] env[65522]: WARNING openstack [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.583048] env[65522]: WARNING openstack [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.629387] env[65522]: DEBUG oslo_concurrency.lockutils [None req-7934ed47-27b9-4c87-9ed6-d021a3fd28ef tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.286s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.633478] env[65522]: DEBUG oslo_concurrency.lockutils [None req-eab79372-f48d-4fb3-839f-6b68f0044ce6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.081s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1109.633724] env[65522]: DEBUG nova.compute.manager [None req-eab79372-f48d-4fb3-839f-6b68f0044ce6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1109.637019] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c99342a-0104-4913-993f-3e9e961cabe9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.644952] env[65522]: DEBUG nova.compute.manager [None req-eab79372-f48d-4fb3-839f-6b68f0044ce6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65522) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1109.645654] env[65522]: DEBUG nova.objects.instance [None req-eab79372-f48d-4fb3-839f-6b68f0044ce6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lazy-loading 'flavor' on Instance uuid 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1109.683166] env[65522]: DEBUG nova.compute.manager [None req-6c3a86cb-baf4-48dd-91f2-54560de66cf5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1109.683456] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6c3a86cb-baf4-48dd-91f2-54560de66cf5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1109.684026] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f41b84d-bbef-40f8-bd3f-b860276fd410 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.695664] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f05619-aecf-45ed-96ec-7a5d664bd21e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.723205] env[65522]: WARNING openstack [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.723656] env[65522]: WARNING openstack [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.747835] env[65522]: WARNING nova.virt.vmwareapi.vmops [None req-6c3a86cb-baf4-48dd-91f2-54560de66cf5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 821a5847-c255-4846-9041-2a0144f4539b could not be found. [ 1109.748060] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6c3a86cb-baf4-48dd-91f2-54560de66cf5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1109.748248] env[65522]: INFO nova.compute.manager [None req-6c3a86cb-baf4-48dd-91f2-54560de66cf5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Took 0.06 seconds to destroy the instance on the hypervisor. [ 1109.748507] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-6c3a86cb-baf4-48dd-91f2-54560de66cf5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1109.748996] env[65522]: DEBUG nova.compute.manager [-] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1109.749148] env[65522]: DEBUG nova.network.neutron [-] [instance: 821a5847-c255-4846-9041-2a0144f4539b] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1109.749352] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1109.749858] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.750164] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.766875] env[65522]: DEBUG oslo_vmware.api [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114811, 'name': RemoveSnapshot_Task, 'duration_secs': 0.675736} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.767510] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1109.814536] env[65522]: WARNING neutronclient.v2_0.client [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1109.815212] env[65522]: WARNING openstack [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1109.815567] env[65522]: WARNING openstack [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1109.833462] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114810, 'name': CloneVM_Task, 'duration_secs': 1.632176} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.834368] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1109.836155] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Created linked-clone VM from snapshot [ 1109.837117] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd410ae-16c9-4b9e-a51c-57415eb24008 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.849016] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Uploading image eb6f13f1-eccd-457d-8ea5-13ff73ac38ae {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1109.879722] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6daba307-2fca-4ac2-b475-3f2ed8c77459 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.884810] env[65522]: DEBUG oslo_vmware.rw_handles [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1109.884810] env[65522]: value = "vm-994935" [ 1109.884810] env[65522]: _type = "VirtualMachine" [ 1109.884810] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1109.885393] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4f6bab9d-9322-4dc4-a525-4011f5950047 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.894610] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd31f74-a597-4397-9aa2-4d6b93568ef5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.898374] env[65522]: DEBUG oslo_vmware.rw_handles [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lease: (returnval){ [ 1109.898374] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52df1034-3353-b6b6-c50b-fbfc813a13b4" [ 1109.898374] env[65522]: _type = "HttpNfcLease" [ 1109.898374] env[65522]: } obtained for exporting VM: (result){ [ 1109.898374] env[65522]: value = "vm-994935" [ 1109.898374] env[65522]: _type = "VirtualMachine" [ 1109.898374] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1109.898735] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the lease: (returnval){ [ 1109.898735] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52df1034-3353-b6b6-c50b-fbfc813a13b4" [ 1109.898735] env[65522]: _type = "HttpNfcLease" [ 1109.898735] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1109.929788] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27a5c9c-9c99-456c-880e-2430284f0cf7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.934355] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1109.934355] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52df1034-3353-b6b6-c50b-fbfc813a13b4" [ 1109.934355] env[65522]: _type = "HttpNfcLease" [ 1109.934355] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1109.935072] env[65522]: DEBUG oslo_vmware.rw_handles [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1109.935072] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52df1034-3353-b6b6-c50b-fbfc813a13b4" [ 1109.935072] env[65522]: _type = "HttpNfcLease" [ 1109.935072] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1109.935869] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c38fdef-a56c-486c-901d-a038a2356502 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.942093] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-410a31bb-8a13-4c84-adb8-58f22e1836c7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.948244] env[65522]: DEBUG nova.network.neutron [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updated VIF entry in instance network info cache for port a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1109.948680] env[65522]: DEBUG nova.network.neutron [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance_info_cache with network_info: [{"id": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "address": "fa:16:3e:5b:f2:d8", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7fc03fa-06", "ovs_interfaceid": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1109.955112] env[65522]: DEBUG oslo_vmware.rw_handles [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b30196-ab39-4f66-5b15-6e4db7aa508f/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1109.955295] env[65522]: DEBUG oslo_vmware.rw_handles [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b30196-ab39-4f66-5b15-6e4db7aa508f/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1110.024389] env[65522]: DEBUG nova.compute.provider_tree [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.087691] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-dc5331a3-bca7-416c-9521-9314092c09a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.271380] env[65522]: WARNING nova.compute.manager [None req-7187ec86-d468-421d-88b6-1fe23a8a2c60 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Image not found during snapshot: nova.exception.ImageNotFound: Image fd2d6661-5057-4e8e-82c2-1523528553b2 could not be found. [ 1110.459342] env[65522]: DEBUG oslo_concurrency.lockutils [req-93abf0fd-4571-4073-81b8-4dce898ff8c7 req-dc2900f9-64fa-4e6c-a770-44b4040e2e5a service nova] Releasing lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1110.529459] env[65522]: DEBUG nova.scheduler.client.report [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1110.604485] env[65522]: DEBUG nova.network.neutron [-] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1110.655730] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-eab79372-f48d-4fb3-839f-6b68f0044ce6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1110.656784] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f04a91cc-83dc-47c6-83d2-d00e5f266771 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.665760] env[65522]: DEBUG oslo_vmware.api [None req-eab79372-f48d-4fb3-839f-6b68f0044ce6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1110.665760] env[65522]: value = "task-5114813" [ 1110.665760] env[65522]: _type = "Task" [ 1110.665760] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.678554] env[65522]: DEBUG oslo_vmware.api [None req-eab79372-f48d-4fb3-839f-6b68f0044ce6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114813, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.035934] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1111.036570] env[65522]: DEBUG nova.compute.manager [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1111.090391] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "45b4a56e-71b8-42ec-b925-a150e667b223" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1111.090702] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "45b4a56e-71b8-42ec-b925-a150e667b223" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1111.090883] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "45b4a56e-71b8-42ec-b925-a150e667b223-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1111.091663] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "45b4a56e-71b8-42ec-b925-a150e667b223-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1111.091852] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "45b4a56e-71b8-42ec-b925-a150e667b223-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1111.094485] env[65522]: INFO nova.compute.manager [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Terminating instance [ 1111.109757] env[65522]: INFO nova.compute.manager [-] [instance: 821a5847-c255-4846-9041-2a0144f4539b] Took 1.36 seconds to deallocate network for instance. [ 1111.186706] env[65522]: DEBUG oslo_vmware.api [None req-eab79372-f48d-4fb3-839f-6b68f0044ce6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114813, 'name': PowerOffVM_Task, 'duration_secs': 0.257829} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.187481] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-eab79372-f48d-4fb3-839f-6b68f0044ce6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1111.187842] env[65522]: DEBUG nova.compute.manager [None req-eab79372-f48d-4fb3-839f-6b68f0044ce6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1111.188885] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1900a0-7a4a-4e53-b19b-1e8e0801a77d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.543049] env[65522]: DEBUG nova.compute.utils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1111.544817] env[65522]: DEBUG nova.compute.manager [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1111.545105] env[65522]: DEBUG nova.network.neutron [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1111.545623] env[65522]: WARNING neutronclient.v2_0.client [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1111.545994] env[65522]: WARNING neutronclient.v2_0.client [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1111.546637] env[65522]: WARNING openstack [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1111.547165] env[65522]: WARNING openstack [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1111.599910] env[65522]: DEBUG nova.compute.manager [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1111.600387] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1111.601474] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d79d64b0-76c8-495b-af8c-17a988336fe8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.608534] env[65522]: DEBUG nova.policy [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a97b2f232f7844c993dccdc386ce3b2c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9614f18e01594bf083e619c11760fe36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1111.619448] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1111.620833] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d4b3f67a-6428-4572-8e76-3be2175581d2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.633634] env[65522]: DEBUG oslo_vmware.api [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1111.633634] env[65522]: value = "task-5114814" [ 1111.633634] env[65522]: _type = "Task" [ 1111.633634] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.650169] env[65522]: DEBUG oslo_vmware.api [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114814, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.703342] env[65522]: DEBUG oslo_concurrency.lockutils [None req-eab79372-f48d-4fb3-839f-6b68f0044ce6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.070s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1111.952495] env[65522]: DEBUG nova.network.neutron [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Successfully created port: 25e1af96-8ed0-4b93-837f-36d7baa1758d {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1112.056076] env[65522]: DEBUG nova.compute.manager [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1112.151701] env[65522]: DEBUG oslo_vmware.api [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114814, 'name': PowerOffVM_Task, 'duration_secs': 0.262349} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.153266] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6c3a86cb-baf4-48dd-91f2-54560de66cf5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "821a5847-c255-4846-9041-2a0144f4539b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.982s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.153872] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1112.154184] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1112.155386] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ceb31143-0f7e-4476-92ae-a3910aa2503f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.247290] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1112.247290] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1112.247290] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleting the datastore file [datastore1] 45b4a56e-71b8-42ec-b925-a150e667b223 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1112.247290] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b81bd35-9fe2-4258-ac78-5a7b768d7a92 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.257112] env[65522]: DEBUG oslo_vmware.api [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for the task: (returnval){ [ 1112.257112] env[65522]: value = "task-5114816" [ 1112.257112] env[65522]: _type = "Task" [ 1112.257112] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.269579] env[65522]: DEBUG oslo_vmware.api [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114816, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.477711] env[65522]: DEBUG nova.objects.instance [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lazy-loading 'flavor' on Instance uuid 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.771199] env[65522]: DEBUG oslo_vmware.api [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Task: {'id': task-5114816, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.288157} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.771814] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1112.772224] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1112.772559] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1112.772871] env[65522]: INFO nova.compute.manager [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1112.773992] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1112.773992] env[65522]: DEBUG nova.compute.manager [-] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1112.773992] env[65522]: DEBUG nova.network.neutron [-] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1112.774175] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1112.774655] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1112.774928] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1112.983869] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.984085] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquired lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1112.984341] env[65522]: DEBUG nova.network.neutron [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1112.984547] env[65522]: DEBUG nova.objects.instance [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lazy-loading 'info_cache' on Instance uuid 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.987670] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1113.067569] env[65522]: DEBUG nova.compute.manager [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1113.103421] env[65522]: DEBUG nova.virt.hardware [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1113.105365] env[65522]: DEBUG nova.virt.hardware [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1113.109018] env[65522]: DEBUG nova.virt.hardware [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1113.109018] env[65522]: DEBUG nova.virt.hardware [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1113.109018] env[65522]: DEBUG nova.virt.hardware [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1113.109018] env[65522]: DEBUG nova.virt.hardware [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1113.109018] env[65522]: DEBUG nova.virt.hardware [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1113.109018] env[65522]: DEBUG nova.virt.hardware [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1113.109018] env[65522]: DEBUG nova.virt.hardware [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1113.109018] env[65522]: DEBUG nova.virt.hardware [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1113.109018] env[65522]: DEBUG nova.virt.hardware [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1113.109018] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed559a0f-c63e-4a9c-b362-3c589b80ee6e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.123755] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eae59e91-1045-4c0c-888e-cf8cfda2cac1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.493901] env[65522]: DEBUG nova.objects.base [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Object Instance<5ce4a286-efd1-4bbc-a23b-931c6701cfe4> lazy-loaded attributes: flavor,info_cache {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1113.691106] env[65522]: DEBUG nova.network.neutron [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Successfully updated port: 25e1af96-8ed0-4b93-837f-36d7baa1758d {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1113.794809] env[65522]: DEBUG nova.network.neutron [-] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1113.876779] env[65522]: DEBUG nova.compute.manager [req-33862ddd-a7ef-402a-abe5-c7788c0b3516 req-58c1ed11-999a-42cd-b0f2-c25b82af6bea service nova] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Received event network-vif-deleted-6fd7e155-ae8c-4b18-8bfc-f3339f1d9e43 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1113.899936] env[65522]: DEBUG nova.compute.manager [req-4ca8c3f6-0921-4295-b604-7a5379446174 req-bb888943-489e-4dd1-9507-d4dd9518438d service nova] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Received event network-vif-plugged-25e1af96-8ed0-4b93-837f-36d7baa1758d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1113.899936] env[65522]: DEBUG oslo_concurrency.lockutils [req-4ca8c3f6-0921-4295-b604-7a5379446174 req-bb888943-489e-4dd1-9507-d4dd9518438d service nova] Acquiring lock "2b61302b-7ce1-4e8d-b981-8d4376d76c3c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1113.900149] env[65522]: DEBUG oslo_concurrency.lockutils [req-4ca8c3f6-0921-4295-b604-7a5379446174 req-bb888943-489e-4dd1-9507-d4dd9518438d service nova] Lock "2b61302b-7ce1-4e8d-b981-8d4376d76c3c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1113.901347] env[65522]: DEBUG oslo_concurrency.lockutils [req-4ca8c3f6-0921-4295-b604-7a5379446174 req-bb888943-489e-4dd1-9507-d4dd9518438d service nova] Lock "2b61302b-7ce1-4e8d-b981-8d4376d76c3c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1113.901347] env[65522]: DEBUG nova.compute.manager [req-4ca8c3f6-0921-4295-b604-7a5379446174 req-bb888943-489e-4dd1-9507-d4dd9518438d service nova] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] No waiting events found dispatching network-vif-plugged-25e1af96-8ed0-4b93-837f-36d7baa1758d {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1113.901347] env[65522]: WARNING nova.compute.manager [req-4ca8c3f6-0921-4295-b604-7a5379446174 req-bb888943-489e-4dd1-9507-d4dd9518438d service nova] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Received unexpected event network-vif-plugged-25e1af96-8ed0-4b93-837f-36d7baa1758d for instance with vm_state building and task_state spawning. [ 1113.999083] env[65522]: WARNING neutronclient.v2_0.client [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1113.999083] env[65522]: WARNING openstack [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1113.999083] env[65522]: WARNING openstack [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.145781] env[65522]: WARNING openstack [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.147473] env[65522]: WARNING openstack [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.199209] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "refresh_cache-2b61302b-7ce1-4e8d-b981-8d4376d76c3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.199209] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "refresh_cache-2b61302b-7ce1-4e8d-b981-8d4376d76c3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1114.199209] env[65522]: DEBUG nova.network.neutron [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1114.217168] env[65522]: WARNING neutronclient.v2_0.client [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.217907] env[65522]: WARNING openstack [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.218163] env[65522]: WARNING openstack [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.297405] env[65522]: INFO nova.compute.manager [-] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Took 1.52 seconds to deallocate network for instance. [ 1114.310195] env[65522]: DEBUG nova.network.neutron [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Updating instance_info_cache with network_info: [{"id": "c2eab4eb-5a45-4c96-a145-bea8b314c82d", "address": "fa:16:3e:22:77:b4", "network": {"id": "f49e9b61-6927-4a5b-a12d-f54f408d42b4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1397158171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a80f979f3dc0477e9462b47f7aa87f14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2eab4eb-5a", "ovs_interfaceid": "c2eab4eb-5a45-4c96-a145-bea8b314c82d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1114.497062] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "3d049de4-4005-4995-a170-5cd6bbf0744c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1114.497476] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "3d049de4-4005-4995-a170-5cd6bbf0744c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1114.527633] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Acquiring lock "95896228-9a70-4f49-bbc9-8f284a666ba9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1114.527945] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Lock "95896228-9a70-4f49-bbc9-8f284a666ba9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1114.699038] env[65522]: WARNING openstack [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.700563] env[65522]: WARNING openstack [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.741061] env[65522]: DEBUG nova.network.neutron [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1114.761819] env[65522]: WARNING openstack [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.762350] env[65522]: WARNING openstack [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.806673] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1114.806946] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1114.807190] env[65522]: DEBUG nova.objects.instance [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lazy-loading 'resources' on Instance uuid 45b4a56e-71b8-42ec-b925-a150e667b223 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1114.813529] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Releasing lock "refresh_cache-5ce4a286-efd1-4bbc-a23b-931c6701cfe4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1114.841560] env[65522]: WARNING neutronclient.v2_0.client [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1114.842336] env[65522]: WARNING openstack [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1114.842699] env[65522]: WARNING openstack [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1114.938625] env[65522]: DEBUG nova.network.neutron [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Updating instance_info_cache with network_info: [{"id": "25e1af96-8ed0-4b93-837f-36d7baa1758d", "address": "fa:16:3e:c4:ef:0b", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e1af96-8e", "ovs_interfaceid": "25e1af96-8ed0-4b93-837f-36d7baa1758d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1115.000958] env[65522]: DEBUG nova.compute.manager [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1115.031239] env[65522]: DEBUG nova.compute.manager [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1115.442487] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "refresh_cache-2b61302b-7ce1-4e8d-b981-8d4376d76c3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1115.442878] env[65522]: DEBUG nova.compute.manager [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Instance network_info: |[{"id": "25e1af96-8ed0-4b93-837f-36d7baa1758d", "address": "fa:16:3e:c4:ef:0b", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e1af96-8e", "ovs_interfaceid": "25e1af96-8ed0-4b93-837f-36d7baa1758d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1115.443690] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:ef:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10b81051-1eb1-406b-888c-4548c470c77e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '25e1af96-8ed0-4b93-837f-36d7baa1758d', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1115.452842] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1115.452842] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1115.452842] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-05cc2735-18f1-4676-b30b-401a0fd103e4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.478699] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1115.478699] env[65522]: value = "task-5114817" [ 1115.478699] env[65522]: _type = "Task" [ 1115.478699] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.490200] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114817, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.525605] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a6584c-4d29-4d9b-bfd2-795034c14dc1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.529862] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1115.540030] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bfffd45-65e2-4c93-89e5-995523b95763 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.579529] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1115.580486] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fbaf03-ebad-4041-a78b-2429f802bcaa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.590251] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1405a4d3-e8dc-4087-b7a9-c82204c94af7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.608684] env[65522]: DEBUG nova.compute.provider_tree [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1115.819853] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1115.820162] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eb26445d-4b6c-4d65-b563-d9ec3d5966b4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.828041] env[65522]: DEBUG oslo_vmware.api [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1115.828041] env[65522]: value = "task-5114818" [ 1115.828041] env[65522]: _type = "Task" [ 1115.828041] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.837318] env[65522]: DEBUG oslo_vmware.api [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114818, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.943413] env[65522]: DEBUG nova.compute.manager [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Received event network-changed-25e1af96-8ed0-4b93-837f-36d7baa1758d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1115.943881] env[65522]: DEBUG nova.compute.manager [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Refreshing instance network info cache due to event network-changed-25e1af96-8ed0-4b93-837f-36d7baa1758d. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1115.946448] env[65522]: DEBUG oslo_concurrency.lockutils [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] Acquiring lock "refresh_cache-2b61302b-7ce1-4e8d-b981-8d4376d76c3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.946448] env[65522]: DEBUG oslo_concurrency.lockutils [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] Acquired lock "refresh_cache-2b61302b-7ce1-4e8d-b981-8d4376d76c3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1115.946448] env[65522]: DEBUG nova.network.neutron [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Refreshing network info cache for port 25e1af96-8ed0-4b93-837f-36d7baa1758d {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1115.990548] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114817, 'name': CreateVM_Task, 'duration_secs': 0.455597} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.990548] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1115.990548] env[65522]: WARNING neutronclient.v2_0.client [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1115.991084] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.991295] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1116.022166] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1116.022166] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8785ee1-6f39-4817-a0b7-8944ae056814 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.022166] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1116.022166] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524cc066-d4c4-64ad-06a0-3ec7452bcdf7" [ 1116.022166] env[65522]: _type = "Task" [ 1116.022166] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.022166] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524cc066-d4c4-64ad-06a0-3ec7452bcdf7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.112542] env[65522]: DEBUG nova.scheduler.client.report [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1116.341455] env[65522]: DEBUG oslo_vmware.api [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114818, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.448879] env[65522]: WARNING neutronclient.v2_0.client [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1116.449659] env[65522]: WARNING openstack [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1116.450046] env[65522]: WARNING openstack [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1116.508196] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524cc066-d4c4-64ad-06a0-3ec7452bcdf7, 'name': SearchDatastore_Task, 'duration_secs': 0.013256} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.508541] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1116.508804] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1116.509064] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.509221] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1116.509404] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1116.509689] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d58cc815-64d9-4dfe-8d96-5404b34f998d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.520013] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1116.520210] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1116.520985] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bce1f545-85b9-4db3-adc0-84670f362ae2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.528076] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1116.528076] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52949e76-947c-6c3c-5921-2c4f76d7558d" [ 1116.528076] env[65522]: _type = "Task" [ 1116.528076] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.542371] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52949e76-947c-6c3c-5921-2c4f76d7558d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.555091] env[65522]: WARNING openstack [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1116.555496] env[65522]: WARNING openstack [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1116.618715] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.812s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1116.622569] env[65522]: WARNING neutronclient.v2_0.client [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1116.623404] env[65522]: WARNING openstack [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1116.623896] env[65522]: WARNING openstack [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1116.632099] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.102s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1116.633839] env[65522]: INFO nova.compute.claims [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1116.658375] env[65522]: INFO nova.scheduler.client.report [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Deleted allocations for instance 45b4a56e-71b8-42ec-b925-a150e667b223 [ 1116.739491] env[65522]: DEBUG nova.network.neutron [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Updated VIF entry in instance network info cache for port 25e1af96-8ed0-4b93-837f-36d7baa1758d. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1116.739880] env[65522]: DEBUG nova.network.neutron [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Updating instance_info_cache with network_info: [{"id": "25e1af96-8ed0-4b93-837f-36d7baa1758d", "address": "fa:16:3e:c4:ef:0b", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e1af96-8e", "ovs_interfaceid": "25e1af96-8ed0-4b93-837f-36d7baa1758d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1116.843483] env[65522]: DEBUG oslo_vmware.api [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114818, 'name': PowerOnVM_Task, 'duration_secs': 0.563126} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.843773] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1116.844845] env[65522]: DEBUG nova.compute.manager [None req-b6526e3f-cf56-4448-9d59-745be7e6ee35 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1116.844845] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e7b8e6-f4be-4b36-90f4-94df03acc2de {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.039318] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52949e76-947c-6c3c-5921-2c4f76d7558d, 'name': SearchDatastore_Task, 'duration_secs': 0.029097} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.040241] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9787e185-ec14-40c7-a081-960818b9fb73 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.046147] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1117.046147] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520ae81e-395f-a16b-ea5d-62485ee36a08" [ 1117.046147] env[65522]: _type = "Task" [ 1117.046147] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.054388] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520ae81e-395f-a16b-ea5d-62485ee36a08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.165946] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b82f37ff-955c-417b-9fcb-3f20931af5b6 tempest-ImagesTestJSON-1194554300 tempest-ImagesTestJSON-1194554300-project-member] Lock "45b4a56e-71b8-42ec-b925-a150e667b223" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.075s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1117.244587] env[65522]: DEBUG oslo_concurrency.lockutils [req-2207daad-4147-4429-8de6-4e1af5f335bb req-c82356c3-8108-49c1-916c-3bf15176d1c2 service nova] Releasing lock "refresh_cache-2b61302b-7ce1-4e8d-b981-8d4376d76c3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1117.558480] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520ae81e-395f-a16b-ea5d-62485ee36a08, 'name': SearchDatastore_Task, 'duration_secs': 0.032686} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.558871] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1117.559063] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 2b61302b-7ce1-4e8d-b981-8d4376d76c3c/2b61302b-7ce1-4e8d-b981-8d4376d76c3c.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1117.559332] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5091a9f3-3ed4-4069-a5f6-86cdd7a6ca5b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.568307] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1117.568307] env[65522]: value = "task-5114819" [ 1117.568307] env[65522]: _type = "Task" [ 1117.568307] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.579969] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114819, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.797771] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56489e6-ddca-4548-9c6a-76497b47cbb1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.806654] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4757c668-b966-4492-9987-262dc61dae04 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.841503] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b4ec00-16c1-4d61-b599-18f1abd85501 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.849805] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbcc6312-e902-4431-8447-9458b243edb7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.864754] env[65522]: DEBUG nova.compute.provider_tree [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1117.872290] env[65522]: DEBUG oslo_concurrency.lockutils [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1117.873047] env[65522]: DEBUG oslo_concurrency.lockutils [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1118.080742] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114819, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.368269] env[65522]: DEBUG nova.scheduler.client.report [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1118.375741] env[65522]: INFO nova.compute.manager [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Detaching volume fd5bf7dd-1d57-4e86-badc-405c9e3774cc [ 1118.417111] env[65522]: INFO nova.virt.block_device [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Attempting to driver detach volume fd5bf7dd-1d57-4e86-badc-405c9e3774cc from mountpoint /dev/sdb [ 1118.417336] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1118.417520] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994901', 'volume_id': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'name': 'volume-fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '79dd37b5-6b30-48ab-9f00-78214cbd132d', 'attached_at': '', 'detached_at': '', 'volume_id': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'serial': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1118.418580] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71db87bb-88e4-4493-a9f6-5fc3d6f3daa1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.444210] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa29f95-d96c-42a4-b63a-9385c8ad4b99 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.454382] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384cd70d-ed2f-43b7-9e6c-b4ac174f004a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.480774] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42523d65-47f1-4bdb-a1c4-b5e558041fb5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.499728] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The volume has not been displaced from its original location: [datastore1] volume-fd5bf7dd-1d57-4e86-badc-405c9e3774cc/volume-fd5bf7dd-1d57-4e86-badc-405c9e3774cc.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1118.505373] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Reconfiguring VM instance instance-0000004c to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1118.505762] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57eaa571-c91e-4760-89a2-e0c618d91c6a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.526110] env[65522]: DEBUG oslo_vmware.api [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1118.526110] env[65522]: value = "task-5114820" [ 1118.526110] env[65522]: _type = "Task" [ 1118.526110] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.535402] env[65522]: DEBUG oslo_vmware.api [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114820, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.581077] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114819, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.905701} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.581077] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 2b61302b-7ce1-4e8d-b981-8d4376d76c3c/2b61302b-7ce1-4e8d-b981-8d4376d76c3c.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1118.581077] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1118.581610] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d8370efa-776c-45a2-8f7e-cfa7db7b1384 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.589666] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1118.589666] env[65522]: value = "task-5114821" [ 1118.589666] env[65522]: _type = "Task" [ 1118.589666] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.599471] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114821, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.874081] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.242s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1118.874653] env[65522]: DEBUG nova.compute.manager [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1118.877876] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.298s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1118.878696] env[65522]: INFO nova.compute.claims [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1119.039154] env[65522]: DEBUG oslo_vmware.api [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114820, 'name': ReconfigVM_Task, 'duration_secs': 0.260674} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.039820] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Reconfigured VM instance instance-0000004c to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1119.045365] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31a3f03e-527d-401c-9315-48cc5f78f9a4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.061965] env[65522]: DEBUG oslo_vmware.api [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1119.061965] env[65522]: value = "task-5114822" [ 1119.061965] env[65522]: _type = "Task" [ 1119.061965] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.072499] env[65522]: DEBUG oslo_vmware.api [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114822, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.100767] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114821, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079595} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.100995] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1119.102070] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8864ec76-b6ec-4bed-b155-6394bed1f275 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.126747] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 2b61302b-7ce1-4e8d-b981-8d4376d76c3c/2b61302b-7ce1-4e8d-b981-8d4376d76c3c.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1119.127145] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d57807f-d852-4154-a8d2-6e9b4618e251 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.150055] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1119.150055] env[65522]: value = "task-5114823" [ 1119.150055] env[65522]: _type = "Task" [ 1119.150055] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.160903] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114823, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.383462] env[65522]: DEBUG nova.compute.utils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1119.387179] env[65522]: DEBUG nova.compute.manager [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1119.387385] env[65522]: DEBUG nova.network.neutron [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1119.387730] env[65522]: WARNING neutronclient.v2_0.client [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1119.388054] env[65522]: WARNING neutronclient.v2_0.client [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1119.388760] env[65522]: WARNING openstack [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1119.389269] env[65522]: WARNING openstack [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1119.448201] env[65522]: DEBUG nova.policy [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8f8fa586d8c74bb9805a1148fa7ed037', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f93c790ff61543bd8e134bcf9cb20bb2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1119.574865] env[65522]: DEBUG oslo_vmware.api [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114822, 'name': ReconfigVM_Task, 'duration_secs': 0.145314} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.575493] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994901', 'volume_id': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'name': 'volume-fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '79dd37b5-6b30-48ab-9f00-78214cbd132d', 'attached_at': '', 'detached_at': '', 'volume_id': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc', 'serial': 'fd5bf7dd-1d57-4e86-badc-405c9e3774cc'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1119.661302] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114823, 'name': ReconfigVM_Task, 'duration_secs': 0.286762} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.661636] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 2b61302b-7ce1-4e8d-b981-8d4376d76c3c/2b61302b-7ce1-4e8d-b981-8d4376d76c3c.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1119.662296] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2b2d7ce5-77ed-42c4-b3a4-42ad38b342bd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.669323] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1119.669323] env[65522]: value = "task-5114824" [ 1119.669323] env[65522]: _type = "Task" [ 1119.669323] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.682184] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114824, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.812840] env[65522]: DEBUG nova.network.neutron [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Successfully created port: 301397ef-592e-4cf7-b2e2-53586abbec6d {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1119.902879] env[65522]: DEBUG nova.compute.manager [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1120.078651] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9fc4e90-0b96-47ff-8b70-65d5f8a88e02 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.090596] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5653a9f-f375-49cc-a50d-ed4d89fac435 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.122080] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556e5f4f-3882-43ed-96cb-90b65b4540a3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.131119] env[65522]: DEBUG nova.objects.instance [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lazy-loading 'flavor' on Instance uuid 79dd37b5-6b30-48ab-9f00-78214cbd132d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.133499] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc780e8-497d-45af-b512-e50e9a181cdb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.151664] env[65522]: DEBUG nova.compute.provider_tree [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.180282] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114824, 'name': Rename_Task, 'duration_secs': 0.151079} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.180642] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1120.180870] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12346072-7dac-4160-8564-d1aa82cd4033 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.189088] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1120.189088] env[65522]: value = "task-5114825" [ 1120.189088] env[65522]: _type = "Task" [ 1120.189088] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.198711] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114825, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.654094] env[65522]: DEBUG nova.scheduler.client.report [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1120.701840] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114825, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.911937] env[65522]: DEBUG nova.compute.manager [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1120.940913] env[65522]: DEBUG nova.virt.hardware [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1120.941167] env[65522]: DEBUG nova.virt.hardware [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1120.941326] env[65522]: DEBUG nova.virt.hardware [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1120.941503] env[65522]: DEBUG nova.virt.hardware [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1120.941640] env[65522]: DEBUG nova.virt.hardware [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1120.941780] env[65522]: DEBUG nova.virt.hardware [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1120.941983] env[65522]: DEBUG nova.virt.hardware [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1120.942156] env[65522]: DEBUG nova.virt.hardware [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1120.942368] env[65522]: DEBUG nova.virt.hardware [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1120.942606] env[65522]: DEBUG nova.virt.hardware [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1120.942811] env[65522]: DEBUG nova.virt.hardware [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1120.943841] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed6d97a-8375-430e-941d-b2b931cefded {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.952671] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd4a995-3a63-459d-8d2b-d70f161fc035 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.143657] env[65522]: DEBUG oslo_concurrency.lockutils [None req-67aca8bc-d8f4-45f1-adef-bfe56f25a65d tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.271s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1121.160107] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1121.160726] env[65522]: DEBUG nova.compute.manager [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1121.184588] env[65522]: DEBUG oslo_vmware.rw_handles [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b30196-ab39-4f66-5b15-6e4db7aa508f/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1121.187075] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5221bc7f-bf7f-4db4-ac73-7c5caf7bf045 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.196246] env[65522]: DEBUG oslo_vmware.rw_handles [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b30196-ab39-4f66-5b15-6e4db7aa508f/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1121.197264] env[65522]: ERROR oslo_vmware.rw_handles [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b30196-ab39-4f66-5b15-6e4db7aa508f/disk-0.vmdk due to incomplete transfer. [ 1121.197264] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2d9f3df1-cb96-419a-9dd2-f2ba51dac6a1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.205248] env[65522]: DEBUG oslo_vmware.api [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114825, 'name': PowerOnVM_Task, 'duration_secs': 0.704803} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.205248] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1121.205248] env[65522]: INFO nova.compute.manager [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Took 8.14 seconds to spawn the instance on the hypervisor. [ 1121.205248] env[65522]: DEBUG nova.compute.manager [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1121.205248] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce033fc-29b1-44c1-8bed-7fcd08d0df5d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.208517] env[65522]: DEBUG oslo_vmware.rw_handles [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52b30196-ab39-4f66-5b15-6e4db7aa508f/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1121.208695] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Uploaded image eb6f13f1-eccd-457d-8ea5-13ff73ac38ae to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1121.211311] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1121.211955] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c50ad1f6-2ede-4dc7-9ea5-86556ea61704 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.220874] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1121.220874] env[65522]: value = "task-5114826" [ 1121.220874] env[65522]: _type = "Task" [ 1121.220874] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.233069] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114826, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.337060] env[65522]: DEBUG nova.network.neutron [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Successfully updated port: 301397ef-592e-4cf7-b2e2-53586abbec6d {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1121.538051] env[65522]: DEBUG nova.compute.manager [req-3b0c2a16-8aca-4093-9537-594075e440f2 req-fdf19097-d148-4177-9b4b-d7e84f997ff8 service nova] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Received event network-vif-plugged-301397ef-592e-4cf7-b2e2-53586abbec6d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1121.538290] env[65522]: DEBUG oslo_concurrency.lockutils [req-3b0c2a16-8aca-4093-9537-594075e440f2 req-fdf19097-d148-4177-9b4b-d7e84f997ff8 service nova] Acquiring lock "3d049de4-4005-4995-a170-5cd6bbf0744c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1121.538560] env[65522]: DEBUG oslo_concurrency.lockutils [req-3b0c2a16-8aca-4093-9537-594075e440f2 req-fdf19097-d148-4177-9b4b-d7e84f997ff8 service nova] Lock "3d049de4-4005-4995-a170-5cd6bbf0744c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1121.538768] env[65522]: DEBUG oslo_concurrency.lockutils [req-3b0c2a16-8aca-4093-9537-594075e440f2 req-fdf19097-d148-4177-9b4b-d7e84f997ff8 service nova] Lock "3d049de4-4005-4995-a170-5cd6bbf0744c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1121.538940] env[65522]: DEBUG nova.compute.manager [req-3b0c2a16-8aca-4093-9537-594075e440f2 req-fdf19097-d148-4177-9b4b-d7e84f997ff8 service nova] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] No waiting events found dispatching network-vif-plugged-301397ef-592e-4cf7-b2e2-53586abbec6d {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1121.539309] env[65522]: WARNING nova.compute.manager [req-3b0c2a16-8aca-4093-9537-594075e440f2 req-fdf19097-d148-4177-9b4b-d7e84f997ff8 service nova] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Received unexpected event network-vif-plugged-301397ef-592e-4cf7-b2e2-53586abbec6d for instance with vm_state building and task_state spawning. [ 1121.667074] env[65522]: DEBUG nova.compute.utils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1121.667219] env[65522]: DEBUG nova.compute.manager [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1121.668082] env[65522]: DEBUG nova.network.neutron [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1121.668082] env[65522]: WARNING neutronclient.v2_0.client [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.668550] env[65522]: WARNING neutronclient.v2_0.client [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.669351] env[65522]: WARNING openstack [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.669689] env[65522]: WARNING openstack [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.679415] env[65522]: DEBUG nova.compute.manager [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1121.729675] env[65522]: INFO nova.compute.manager [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Took 14.59 seconds to build instance. [ 1121.744888] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114826, 'name': Destroy_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.808587] env[65522]: DEBUG nova.policy [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05a25f6e13f046858277100f442b6bf6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eaac83f5267c4f89a8d2829d5eb32af8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1121.840895] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "refresh_cache-3d049de4-4005-4995-a170-5cd6bbf0744c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.841106] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "refresh_cache-3d049de4-4005-4995-a170-5cd6bbf0744c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1121.841297] env[65522]: DEBUG nova.network.neutron [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1122.142970] env[65522]: DEBUG oslo_concurrency.lockutils [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1122.144383] env[65522]: DEBUG oslo_concurrency.lockutils [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1122.144383] env[65522]: DEBUG oslo_concurrency.lockutils [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "79dd37b5-6b30-48ab-9f00-78214cbd132d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1122.144383] env[65522]: DEBUG oslo_concurrency.lockutils [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1122.144383] env[65522]: DEBUG oslo_concurrency.lockutils [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1122.147619] env[65522]: INFO nova.compute.manager [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Terminating instance [ 1122.212946] env[65522]: DEBUG nova.network.neutron [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Successfully created port: 672d56ee-9ded-4c3a-971d-e807052d71c8 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1122.232135] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cb7d6672-b7d4-44fb-a9ee-2e87a2d7beed tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "2b61302b-7ce1-4e8d-b981-8d4376d76c3c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.102s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1122.246522] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114826, 'name': Destroy_Task, 'duration_secs': 0.967092} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.246770] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Destroyed the VM [ 1122.247024] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1122.247479] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7c528bb2-7a6b-4e13-a1d9-3818f80452b3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.255614] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1122.255614] env[65522]: value = "task-5114827" [ 1122.255614] env[65522]: _type = "Task" [ 1122.255614] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.266260] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114827, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.349586] env[65522]: WARNING openstack [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.350381] env[65522]: WARNING openstack [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.416745] env[65522]: DEBUG nova.network.neutron [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1122.456175] env[65522]: WARNING openstack [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.456626] env[65522]: WARNING openstack [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.632017] env[65522]: WARNING neutronclient.v2_0.client [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1122.632782] env[65522]: WARNING openstack [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.633231] env[65522]: WARNING openstack [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.652792] env[65522]: DEBUG nova.compute.manager [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1122.653013] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1122.653984] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5adeb70c-b446-4a78-9f5b-12169ec385fc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.662704] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1122.663114] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e621937d-47d8-4fc2-bd49-1b2505fc9246 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.671759] env[65522]: DEBUG oslo_vmware.api [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1122.671759] env[65522]: value = "task-5114828" [ 1122.671759] env[65522]: _type = "Task" [ 1122.671759] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.681804] env[65522]: DEBUG oslo_vmware.api [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114828, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.692396] env[65522]: DEBUG nova.compute.manager [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1122.723915] env[65522]: DEBUG nova.virt.hardware [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1122.724168] env[65522]: DEBUG nova.virt.hardware [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1122.724325] env[65522]: DEBUG nova.virt.hardware [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1122.724502] env[65522]: DEBUG nova.virt.hardware [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1122.724676] env[65522]: DEBUG nova.virt.hardware [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1122.724829] env[65522]: DEBUG nova.virt.hardware [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1122.725361] env[65522]: DEBUG nova.virt.hardware [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1122.725543] env[65522]: DEBUG nova.virt.hardware [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1122.725895] env[65522]: DEBUG nova.virt.hardware [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1122.726077] env[65522]: DEBUG nova.virt.hardware [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1122.726259] env[65522]: DEBUG nova.virt.hardware [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1122.727333] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ceda455-efc6-447c-bc20-feda1ec86acf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.737149] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0377b047-9187-4fcb-9d75-ad0ec0fd23dd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.768203] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114827, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.826533] env[65522]: INFO nova.compute.manager [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Rebuilding instance [ 1122.880896] env[65522]: DEBUG nova.compute.manager [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1122.881476] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b26fa22-acfe-46d8-83f4-849b74fa0239 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.887325] env[65522]: DEBUG nova.network.neutron [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Updating instance_info_cache with network_info: [{"id": "301397ef-592e-4cf7-b2e2-53586abbec6d", "address": "fa:16:3e:1a:6d:23", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap301397ef-59", "ovs_interfaceid": "301397ef-592e-4cf7-b2e2-53586abbec6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1123.183763] env[65522]: DEBUG oslo_vmware.api [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114828, 'name': PowerOffVM_Task, 'duration_secs': 0.35152} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.184125] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1123.184329] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1123.184694] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7da94a59-4072-4e6c-a1c1-d438dfdf5d35 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.254709] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1123.254967] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1123.255256] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleting the datastore file [datastore2] 79dd37b5-6b30-48ab-9f00-78214cbd132d {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1123.255594] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-706a0c2b-8abe-4949-a06c-f2b2d45d85e2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.272867] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114827, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.274759] env[65522]: DEBUG oslo_vmware.api [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1123.274759] env[65522]: value = "task-5114830" [ 1123.274759] env[65522]: _type = "Task" [ 1123.274759] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.287660] env[65522]: DEBUG oslo_vmware.api [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114830, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.393449] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "refresh_cache-3d049de4-4005-4995-a170-5cd6bbf0744c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1123.393861] env[65522]: DEBUG nova.compute.manager [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Instance network_info: |[{"id": "301397ef-592e-4cf7-b2e2-53586abbec6d", "address": "fa:16:3e:1a:6d:23", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap301397ef-59", "ovs_interfaceid": "301397ef-592e-4cf7-b2e2-53586abbec6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1123.394365] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:6d:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '301397ef-592e-4cf7-b2e2-53586abbec6d', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1123.402709] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1123.405206] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1123.405573] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-54b1f943-af02-4577-8c8a-04ee8ce513af {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.430502] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1123.430502] env[65522]: value = "task-5114831" [ 1123.430502] env[65522]: _type = "Task" [ 1123.430502] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.439994] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114831, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.567376] env[65522]: DEBUG nova.compute.manager [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Received event network-changed-301397ef-592e-4cf7-b2e2-53586abbec6d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1123.567564] env[65522]: DEBUG nova.compute.manager [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Refreshing instance network info cache due to event network-changed-301397ef-592e-4cf7-b2e2-53586abbec6d. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1123.567776] env[65522]: DEBUG oslo_concurrency.lockutils [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] Acquiring lock "refresh_cache-3d049de4-4005-4995-a170-5cd6bbf0744c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.567947] env[65522]: DEBUG oslo_concurrency.lockutils [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] Acquired lock "refresh_cache-3d049de4-4005-4995-a170-5cd6bbf0744c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1123.568278] env[65522]: DEBUG nova.network.neutron [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Refreshing network info cache for port 301397ef-592e-4cf7-b2e2-53586abbec6d {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1123.666812] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1123.667112] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1123.708595] env[65522]: DEBUG nova.compute.manager [req-c423cd97-2d7f-4839-8bab-3298cce5f0d7 req-57b89db8-29de-4ccf-a359-c7d80811f1a0 service nova] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Received event network-vif-plugged-672d56ee-9ded-4c3a-971d-e807052d71c8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1123.708776] env[65522]: DEBUG oslo_concurrency.lockutils [req-c423cd97-2d7f-4839-8bab-3298cce5f0d7 req-57b89db8-29de-4ccf-a359-c7d80811f1a0 service nova] Acquiring lock "95896228-9a70-4f49-bbc9-8f284a666ba9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1123.708973] env[65522]: DEBUG oslo_concurrency.lockutils [req-c423cd97-2d7f-4839-8bab-3298cce5f0d7 req-57b89db8-29de-4ccf-a359-c7d80811f1a0 service nova] Lock "95896228-9a70-4f49-bbc9-8f284a666ba9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1123.709184] env[65522]: DEBUG oslo_concurrency.lockutils [req-c423cd97-2d7f-4839-8bab-3298cce5f0d7 req-57b89db8-29de-4ccf-a359-c7d80811f1a0 service nova] Lock "95896228-9a70-4f49-bbc9-8f284a666ba9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1123.709487] env[65522]: DEBUG nova.compute.manager [req-c423cd97-2d7f-4839-8bab-3298cce5f0d7 req-57b89db8-29de-4ccf-a359-c7d80811f1a0 service nova] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] No waiting events found dispatching network-vif-plugged-672d56ee-9ded-4c3a-971d-e807052d71c8 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1123.710033] env[65522]: WARNING nova.compute.manager [req-c423cd97-2d7f-4839-8bab-3298cce5f0d7 req-57b89db8-29de-4ccf-a359-c7d80811f1a0 service nova] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Received unexpected event network-vif-plugged-672d56ee-9ded-4c3a-971d-e807052d71c8 for instance with vm_state building and task_state spawning. [ 1123.771925] env[65522]: DEBUG oslo_vmware.api [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114827, 'name': RemoveSnapshot_Task, 'duration_secs': 1.151473} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.772256] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1123.772536] env[65522]: INFO nova.compute.manager [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Took 18.07 seconds to snapshot the instance on the hypervisor. [ 1123.785954] env[65522]: DEBUG oslo_vmware.api [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114830, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211651} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.786272] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1123.786460] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1123.786668] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1123.786811] env[65522]: INFO nova.compute.manager [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1123.787138] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1123.787375] env[65522]: DEBUG nova.compute.manager [-] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1123.787487] env[65522]: DEBUG nova.network.neutron [-] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1123.787744] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1123.789139] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.789139] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.812503] env[65522]: DEBUG nova.network.neutron [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Successfully updated port: 672d56ee-9ded-4c3a-971d-e807052d71c8 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1123.830176] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1123.906975] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1123.907373] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cbe554e5-8f41-4a31-ad2b-69c2c5085b13 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.915505] env[65522]: DEBUG oslo_vmware.api [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Waiting for the task: (returnval){ [ 1123.915505] env[65522]: value = "task-5114832" [ 1123.915505] env[65522]: _type = "Task" [ 1123.915505] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.925825] env[65522]: DEBUG oslo_vmware.api [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114832, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.941783] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114831, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.070980] env[65522]: WARNING neutronclient.v2_0.client [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.071964] env[65522]: WARNING openstack [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.072473] env[65522]: WARNING openstack [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.086853] env[65522]: DEBUG oslo_concurrency.lockutils [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "2b61302b-7ce1-4e8d-b981-8d4376d76c3c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1124.087111] env[65522]: DEBUG oslo_concurrency.lockutils [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "2b61302b-7ce1-4e8d-b981-8d4376d76c3c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1124.087320] env[65522]: DEBUG oslo_concurrency.lockutils [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "2b61302b-7ce1-4e8d-b981-8d4376d76c3c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1124.087527] env[65522]: DEBUG oslo_concurrency.lockutils [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "2b61302b-7ce1-4e8d-b981-8d4376d76c3c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1124.087682] env[65522]: DEBUG oslo_concurrency.lockutils [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "2b61302b-7ce1-4e8d-b981-8d4376d76c3c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1124.091192] env[65522]: INFO nova.compute.manager [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Terminating instance [ 1124.169600] env[65522]: DEBUG nova.compute.manager [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1124.308879] env[65522]: WARNING openstack [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.309286] env[65522]: WARNING openstack [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.317588] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Acquiring lock "refresh_cache-95896228-9a70-4f49-bbc9-8f284a666ba9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.318114] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Acquired lock "refresh_cache-95896228-9a70-4f49-bbc9-8f284a666ba9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1124.318114] env[65522]: DEBUG nova.network.neutron [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1124.365684] env[65522]: DEBUG nova.compute.manager [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Found 3 images (rotation: 2) {{(pid=65522) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 1124.365684] env[65522]: DEBUG nova.compute.manager [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Rotating out 1 backups {{(pid=65522) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5082}} [ 1124.365684] env[65522]: DEBUG nova.compute.manager [None req-07dd141c-d85d-46f9-a02d-040c2c7ec1a1 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Deleting image c64abcc1-1c3f-4b83-a6c1-5a8864327b3b {{(pid=65522) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5087}} [ 1124.424526] env[65522]: WARNING neutronclient.v2_0.client [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.425674] env[65522]: WARNING openstack [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.426199] env[65522]: WARNING openstack [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.445030] env[65522]: DEBUG oslo_vmware.api [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114832, 'name': PowerOffVM_Task, 'duration_secs': 0.285647} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.445294] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1124.445847] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1124.446152] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f327dfd7-ed5e-40d9-9f59-b6c27483f223 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.451538] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114831, 'name': CreateVM_Task, 'duration_secs': 0.543848} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.451538] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1124.451999] env[65522]: WARNING neutronclient.v2_0.client [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.452507] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.452667] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1124.452981] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1124.453290] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a605fdbb-e886-4060-91bd-ce3b1ed7cdad {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.457614] env[65522]: DEBUG oslo_vmware.api [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Waiting for the task: (returnval){ [ 1124.457614] env[65522]: value = "task-5114833" [ 1124.457614] env[65522]: _type = "Task" [ 1124.457614] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.459448] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1124.459448] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524910ed-6b31-293f-213d-d73ddf9cc503" [ 1124.459448] env[65522]: _type = "Task" [ 1124.459448] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.484684] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1124.484931] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1124.485387] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994915', 'volume_id': 'b4d96581-08cb-4fb8-98a2-bc1957d04206', 'name': 'volume-b4d96581-08cb-4fb8-98a2-bc1957d04206', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f6980679-bc2e-442f-b424-8cf80de86c02', 'attached_at': '', 'detached_at': '', 'volume_id': 'b4d96581-08cb-4fb8-98a2-bc1957d04206', 'serial': 'b4d96581-08cb-4fb8-98a2-bc1957d04206'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1124.485640] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524910ed-6b31-293f-213d-d73ddf9cc503, 'name': SearchDatastore_Task, 'duration_secs': 0.011993} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.493457] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6790d7-413e-4537-a1b2-27b2902df454 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.496405] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1124.496405] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1124.496729] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.496802] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1124.496912] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1124.497546] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ca5fa39-0fcb-4a87-8971-dd0239d5a3a0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.517826] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad67ef05-7424-4308-ad5e-30024af08cb7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.520660] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1124.520793] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1124.521819] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b4b95d9-cd6f-451a-9c57-ca1bb3863565 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.534902] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1124.534902] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528ba629-0a6c-dca4-43e0-b582e4bd6e71" [ 1124.534902] env[65522]: _type = "Task" [ 1124.534902] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.534902] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c42e183-5440-47b4-959b-9dabc7873e1d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.544716] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528ba629-0a6c-dca4-43e0-b582e4bd6e71, 'name': SearchDatastore_Task, 'duration_secs': 0.010648} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.565475] env[65522]: DEBUG nova.network.neutron [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Updated VIF entry in instance network info cache for port 301397ef-592e-4cf7-b2e2-53586abbec6d. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1124.566094] env[65522]: DEBUG nova.network.neutron [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Updating instance_info_cache with network_info: [{"id": "301397ef-592e-4cf7-b2e2-53586abbec6d", "address": "fa:16:3e:1a:6d:23", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap301397ef-59", "ovs_interfaceid": "301397ef-592e-4cf7-b2e2-53586abbec6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1124.567924] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b365a178-0485-4e21-9add-3f4ee109ff58 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.572359] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499e12f5-c90e-4164-8539-1f951e21d639 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.580492] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1124.580492] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5273c794-607d-2760-a578-cd415eb0339c" [ 1124.580492] env[65522]: _type = "Task" [ 1124.580492] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.591894] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] The volume has not been displaced from its original location: [datastore1] volume-b4d96581-08cb-4fb8-98a2-bc1957d04206/volume-b4d96581-08cb-4fb8-98a2-bc1957d04206.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1124.597255] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Reconfiguring VM instance instance-00000060 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1124.601859] env[65522]: DEBUG nova.compute.manager [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1124.602061] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1124.602326] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3926d219-1abf-4ace-87fa-cec996dc97ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.616790] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5724a765-6560-48dd-bbdd-8e4085972eed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.626505] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5273c794-607d-2760-a578-cd415eb0339c, 'name': SearchDatastore_Task, 'duration_secs': 0.012669} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.629683] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1124.629864] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 3d049de4-4005-4995-a170-5cd6bbf0744c/3d049de4-4005-4995-a170-5cd6bbf0744c.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1124.633321] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1124.633321] env[65522]: DEBUG oslo_vmware.api [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Waiting for the task: (returnval){ [ 1124.633321] env[65522]: value = "task-5114834" [ 1124.633321] env[65522]: _type = "Task" [ 1124.633321] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.633321] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-92312497-08e5-45f8-a52b-4bc1ed126551 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.633321] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e314191-50fa-474c-acdc-b4bf84922558 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.641338] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1124.641338] env[65522]: value = "task-5114835" [ 1124.641338] env[65522]: _type = "Task" [ 1124.641338] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.649067] env[65522]: DEBUG oslo_vmware.api [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114834, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.649457] env[65522]: DEBUG oslo_vmware.api [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1124.649457] env[65522]: value = "task-5114836" [ 1124.649457] env[65522]: _type = "Task" [ 1124.649457] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.656951] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114835, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.662566] env[65522]: DEBUG oslo_vmware.api [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114836, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.698557] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1124.698850] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1124.700466] env[65522]: INFO nova.compute.claims [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1124.821029] env[65522]: WARNING openstack [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.821406] env[65522]: WARNING openstack [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.845100] env[65522]: DEBUG nova.network.neutron [-] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1124.879523] env[65522]: DEBUG nova.network.neutron [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1124.904340] env[65522]: WARNING openstack [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.904759] env[65522]: WARNING openstack [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.981653] env[65522]: WARNING neutronclient.v2_0.client [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.982423] env[65522]: WARNING openstack [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.982787] env[65522]: WARNING openstack [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1125.077466] env[65522]: DEBUG oslo_concurrency.lockutils [req-3b63f052-5e72-4857-982d-1edd5e48b924 req-6e174a63-6d08-4d55-9630-082ab69d72f3 service nova] Releasing lock "refresh_cache-3d049de4-4005-4995-a170-5cd6bbf0744c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1125.095400] env[65522]: DEBUG nova.network.neutron [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Updating instance_info_cache with network_info: [{"id": "672d56ee-9ded-4c3a-971d-e807052d71c8", "address": "fa:16:3e:fa:be:87", "network": {"id": "96ba0f09-2bb9-455d-9746-ed7a76dc50aa", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1035469686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eaac83f5267c4f89a8d2829d5eb32af8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap672d56ee-9d", "ovs_interfaceid": "672d56ee-9ded-4c3a-971d-e807052d71c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1125.148888] env[65522]: DEBUG oslo_vmware.api [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114834, 'name': ReconfigVM_Task, 'duration_secs': 0.212616} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.149204] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Reconfigured VM instance instance-00000060 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1125.154623] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53eb0992-663b-4e78-a60e-5542f7e2114b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.171823] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114835, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.178437] env[65522]: DEBUG oslo_vmware.api [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114836, 'name': PowerOffVM_Task, 'duration_secs': 0.233938} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.179948] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1125.180161] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1125.180495] env[65522]: DEBUG oslo_vmware.api [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Waiting for the task: (returnval){ [ 1125.180495] env[65522]: value = "task-5114837" [ 1125.180495] env[65522]: _type = "Task" [ 1125.180495] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.180715] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a67643b1-8538-4814-a7a4-61e2d4bf5403 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.193378] env[65522]: DEBUG oslo_vmware.api [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114837, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.254328] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1125.254676] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1125.254760] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleting the datastore file [datastore1] 2b61302b-7ce1-4e8d-b981-8d4376d76c3c {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1125.255123] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-688592a5-ca60-4d91-94f0-04343e8e2849 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.263058] env[65522]: DEBUG oslo_vmware.api [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1125.263058] env[65522]: value = "task-5114839" [ 1125.263058] env[65522]: _type = "Task" [ 1125.263058] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.273030] env[65522]: DEBUG oslo_vmware.api [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114839, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.349766] env[65522]: INFO nova.compute.manager [-] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Took 1.56 seconds to deallocate network for instance. [ 1125.600291] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Releasing lock "refresh_cache-95896228-9a70-4f49-bbc9-8f284a666ba9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1125.600291] env[65522]: DEBUG nova.compute.manager [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Instance network_info: |[{"id": "672d56ee-9ded-4c3a-971d-e807052d71c8", "address": "fa:16:3e:fa:be:87", "network": {"id": "96ba0f09-2bb9-455d-9746-ed7a76dc50aa", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1035469686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eaac83f5267c4f89a8d2829d5eb32af8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap672d56ee-9d", "ovs_interfaceid": "672d56ee-9ded-4c3a-971d-e807052d71c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1125.600291] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:be:87', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db068f71-08cc-42d4-8ab6-17134c1585e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '672d56ee-9ded-4c3a-971d-e807052d71c8', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1125.609828] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Creating folder: Project (eaac83f5267c4f89a8d2829d5eb32af8). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1125.609828] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-958b1f94-42d4-458d-ba10-a372c8e21b17 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.622181] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Created folder: Project (eaac83f5267c4f89a8d2829d5eb32af8) in parent group-v994660. [ 1125.622414] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Creating folder: Instances. Parent ref: group-v994938. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1125.622690] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a13d5c17-3e51-4254-83a5-8f2ecb74db09 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.633334] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Created folder: Instances in parent group-v994938. [ 1125.633611] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1125.633839] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1125.634075] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2b8f5031-fdcd-4fad-bca9-f844faecab75 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.660217] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114835, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534699} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.661540] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 3d049de4-4005-4995-a170-5cd6bbf0744c/3d049de4-4005-4995-a170-5cd6bbf0744c.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1125.661904] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1125.662386] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1125.662386] env[65522]: value = "task-5114842" [ 1125.662386] env[65522]: _type = "Task" [ 1125.662386] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.662519] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2a4557d-68e8-49d8-b999-a98e376c1ff8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.672353] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114842, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.673692] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1125.673692] env[65522]: value = "task-5114843" [ 1125.673692] env[65522]: _type = "Task" [ 1125.673692] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.682020] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114843, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.692234] env[65522]: DEBUG oslo_vmware.api [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114837, 'name': ReconfigVM_Task, 'duration_secs': 0.203914} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.692564] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994915', 'volume_id': 'b4d96581-08cb-4fb8-98a2-bc1957d04206', 'name': 'volume-b4d96581-08cb-4fb8-98a2-bc1957d04206', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f6980679-bc2e-442f-b424-8cf80de86c02', 'attached_at': '', 'detached_at': '', 'volume_id': 'b4d96581-08cb-4fb8-98a2-bc1957d04206', 'serial': 'b4d96581-08cb-4fb8-98a2-bc1957d04206'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1125.692848] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1125.693704] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beafc1fd-a315-4e12-8210-d3aa636c412a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.701268] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1125.701538] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc517c31-21ac-45da-9016-691e1f377f34 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.746914] env[65522]: DEBUG nova.compute.manager [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Received event network-changed-672d56ee-9ded-4c3a-971d-e807052d71c8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1125.746914] env[65522]: DEBUG nova.compute.manager [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Refreshing instance network info cache due to event network-changed-672d56ee-9ded-4c3a-971d-e807052d71c8. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1125.747149] env[65522]: DEBUG oslo_concurrency.lockutils [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] Acquiring lock "refresh_cache-95896228-9a70-4f49-bbc9-8f284a666ba9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.747225] env[65522]: DEBUG oslo_concurrency.lockutils [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] Acquired lock "refresh_cache-95896228-9a70-4f49-bbc9-8f284a666ba9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1125.747430] env[65522]: DEBUG nova.network.neutron [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Refreshing network info cache for port 672d56ee-9ded-4c3a-971d-e807052d71c8 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1125.771284] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1125.771284] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1125.771284] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Deleting the datastore file [datastore1] f6980679-bc2e-442f-b424-8cf80de86c02 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1125.771284] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9afd24ed-ab77-42c9-891a-76247bb16471 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.781213] env[65522]: DEBUG oslo_vmware.api [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114839, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.256112} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.786469] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1125.786469] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1125.786469] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1125.786469] env[65522]: INFO nova.compute.manager [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1125.786469] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1125.786725] env[65522]: DEBUG oslo_vmware.api [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Waiting for the task: (returnval){ [ 1125.786725] env[65522]: value = "task-5114845" [ 1125.786725] env[65522]: _type = "Task" [ 1125.786725] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.786884] env[65522]: DEBUG nova.compute.manager [-] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1125.786994] env[65522]: DEBUG nova.network.neutron [-] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1125.787221] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1125.787829] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1125.788103] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1125.807699] env[65522]: DEBUG oslo_vmware.api [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114845, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.826607] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1125.859309] env[65522]: DEBUG oslo_concurrency.lockutils [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.910618] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6506bb8d-63ed-47a8-88c3-96e0dfdf7990 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.922653] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8950bf64-b915-4e37-b08d-d54015c85053 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.958299] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13080e30-a206-4e7a-99ff-fbaeaccf3306 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.967900] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79866e0-6d6f-41c7-aa22-a32b6769d262 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.984889] env[65522]: DEBUG nova.compute.provider_tree [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1126.076041] env[65522]: DEBUG nova.compute.manager [req-55bae411-07c0-451a-b1a0-e81d882e2a21 req-a98cc908-a687-489a-81a6-54012bd5fc34 service nova] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Received event network-vif-deleted-25e1af96-8ed0-4b93-837f-36d7baa1758d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1126.076227] env[65522]: INFO nova.compute.manager [req-55bae411-07c0-451a-b1a0-e81d882e2a21 req-a98cc908-a687-489a-81a6-54012bd5fc34 service nova] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Neutron deleted interface 25e1af96-8ed0-4b93-837f-36d7baa1758d; detaching it from the instance and deleting it from the info cache [ 1126.076431] env[65522]: DEBUG nova.network.neutron [req-55bae411-07c0-451a-b1a0-e81d882e2a21 req-a98cc908-a687-489a-81a6-54012bd5fc34 service nova] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1126.175359] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114842, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.184680] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114843, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077209} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.184953] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1126.185840] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1792ae5-199a-4bfa-bcb6-728cbf784625 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.209024] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 3d049de4-4005-4995-a170-5cd6bbf0744c/3d049de4-4005-4995-a170-5cd6bbf0744c.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1126.209797] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3f7a577-7cd9-4d1d-834f-67e3c22ec8ab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.229910] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1126.229910] env[65522]: value = "task-5114846" [ 1126.229910] env[65522]: _type = "Task" [ 1126.229910] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.238687] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114846, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.253745] env[65522]: WARNING neutronclient.v2_0.client [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1126.254431] env[65522]: WARNING openstack [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.254796] env[65522]: WARNING openstack [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.304302] env[65522]: DEBUG oslo_vmware.api [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Task: {'id': task-5114845, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.2928} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.304571] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1126.304770] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1126.304952] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1126.358168] env[65522]: WARNING openstack [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.358619] env[65522]: WARNING openstack [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.366795] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1126.367149] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-036f8673-876e-4670-80b9-ab0b71f63cc8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.378198] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d96d59c-fa52-4c8c-81ce-f9386ac36930 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.419538] env[65522]: ERROR nova.compute.manager [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Failed to detach volume b4d96581-08cb-4fb8-98a2-bc1957d04206 from /dev/sda: nova.exception.InstanceNotFound: Instance f6980679-bc2e-442f-b424-8cf80de86c02 could not be found. [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Traceback (most recent call last): [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/compute/manager.py", line 4241, in _do_rebuild_instance [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] self.driver.rebuild(**kwargs) [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/virt/driver.py", line 533, in rebuild [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] raise NotImplementedError() [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] NotImplementedError [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] During handling of the above exception, another exception occurred: [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Traceback (most recent call last): [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/compute/manager.py", line 3664, in _detach_root_volume [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] self.driver.detach_volume(context, old_connection_info, [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] return self._volumeops.detach_volume(connection_info, instance) [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] self._detach_volume_vmdk(connection_info, instance) [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] stable_ref.fetch_moref(session) [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] nova.exception.InstanceNotFound: Instance f6980679-bc2e-442f-b424-8cf80de86c02 could not be found. [ 1126.419538] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] [ 1126.421989] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f0e0969b-ff3b-4ad4-9c31-5a25bddd924c tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1126.422239] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f0e0969b-ff3b-4ad4-9c31-5a25bddd924c tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1126.422437] env[65522]: DEBUG nova.compute.manager [None req-f0e0969b-ff3b-4ad4-9c31-5a25bddd924c tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1126.423362] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47965c3-321c-4cbe-8e5a-3b04ed1c4be9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.427162] env[65522]: WARNING neutronclient.v2_0.client [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1126.427983] env[65522]: WARNING openstack [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.428188] env[65522]: WARNING openstack [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.440779] env[65522]: DEBUG nova.compute.manager [None req-f0e0969b-ff3b-4ad4-9c31-5a25bddd924c tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65522) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1126.441388] env[65522]: DEBUG nova.objects.instance [None req-f0e0969b-ff3b-4ad4-9c31-5a25bddd924c tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'flavor' on Instance uuid 9f342b89-bde2-4c35-ae42-cfe1e6973b74 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1126.487751] env[65522]: DEBUG nova.scheduler.client.report [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1126.525034] env[65522]: DEBUG nova.network.neutron [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Updated VIF entry in instance network info cache for port 672d56ee-9ded-4c3a-971d-e807052d71c8. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1126.525418] env[65522]: DEBUG nova.network.neutron [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Updating instance_info_cache with network_info: [{"id": "672d56ee-9ded-4c3a-971d-e807052d71c8", "address": "fa:16:3e:fa:be:87", "network": {"id": "96ba0f09-2bb9-455d-9746-ed7a76dc50aa", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1035469686-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eaac83f5267c4f89a8d2829d5eb32af8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db068f71-08cc-42d4-8ab6-17134c1585e5", "external-id": "nsx-vlan-transportzone-721", "segmentation_id": 721, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap672d56ee-9d", "ovs_interfaceid": "672d56ee-9ded-4c3a-971d-e807052d71c8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1126.554590] env[65522]: DEBUG nova.network.neutron [-] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1126.579435] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa29172a-0f9a-4136-80cf-6f6ea9ab75b2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.585490] env[65522]: DEBUG nova.compute.utils [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Build of instance f6980679-bc2e-442f-b424-8cf80de86c02 aborted: Failed to rebuild volume backed instance. {{(pid=65522) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 1126.588011] env[65522]: ERROR nova.compute.manager [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance f6980679-bc2e-442f-b424-8cf80de86c02 aborted: Failed to rebuild volume backed instance. [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Traceback (most recent call last): [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/compute/manager.py", line 4241, in _do_rebuild_instance [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] self.driver.rebuild(**kwargs) [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/virt/driver.py", line 533, in rebuild [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] raise NotImplementedError() [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] NotImplementedError [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] During handling of the above exception, another exception occurred: [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Traceback (most recent call last): [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/compute/manager.py", line 3699, in _rebuild_volume_backed_instance [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] self._detach_root_volume(context, instance, root_bdm) [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/compute/manager.py", line 3678, in _detach_root_volume [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] with excutils.save_and_reraise_exception(): [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] self.force_reraise() [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] raise self.value [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/compute/manager.py", line 3664, in _detach_root_volume [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] self.driver.detach_volume(context, old_connection_info, [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] return self._volumeops.detach_volume(connection_info, instance) [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] self._detach_volume_vmdk(connection_info, instance) [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] stable_ref.fetch_moref(session) [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] nova.exception.InstanceNotFound: Instance f6980679-bc2e-442f-b424-8cf80de86c02 could not be found. [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] [ 1126.588011] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] During handling of the above exception, another exception occurred: [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Traceback (most recent call last): [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/compute/manager.py", line 11570, in _error_out_instance_on_exception [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] yield [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/compute/manager.py", line 3967, in rebuild_instance [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] self._do_rebuild_instance_with_claim( [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/compute/manager.py", line 4053, in _do_rebuild_instance_with_claim [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] self._do_rebuild_instance( [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/compute/manager.py", line 4245, in _do_rebuild_instance [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] self._rebuild_default_impl(**kwargs) [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/compute/manager.py", line 3822, in _rebuild_default_impl [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] self._rebuild_volume_backed_instance( [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] File "/opt/stack/nova/nova/compute/manager.py", line 3714, in _rebuild_volume_backed_instance [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] raise exception.BuildAbortException( [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] nova.exception.BuildAbortException: Build of instance f6980679-bc2e-442f-b424-8cf80de86c02 aborted: Failed to rebuild volume backed instance. [ 1126.588894] env[65522]: ERROR nova.compute.manager [instance: f6980679-bc2e-442f-b424-8cf80de86c02] [ 1126.597019] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15ef58f-e6e5-46b3-b0a6-d6438ce50863 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.632501] env[65522]: DEBUG nova.compute.manager [req-55bae411-07c0-451a-b1a0-e81d882e2a21 req-a98cc908-a687-489a-81a6-54012bd5fc34 service nova] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Detach interface failed, port_id=25e1af96-8ed0-4b93-837f-36d7baa1758d, reason: Instance 2b61302b-7ce1-4e8d-b981-8d4376d76c3c could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1126.675635] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114842, 'name': CreateVM_Task, 'duration_secs': 0.589809} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.675854] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1126.676410] env[65522]: WARNING neutronclient.v2_0.client [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1126.676838] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.676997] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1126.677339] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1126.677622] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77d6cef7-01cc-4373-98d3-229efb7d6a4b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.684012] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Waiting for the task: (returnval){ [ 1126.684012] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52aaa09e-e500-75c5-a7d6-3da0ea52af43" [ 1126.684012] env[65522]: _type = "Task" [ 1126.684012] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.693600] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52aaa09e-e500-75c5-a7d6-3da0ea52af43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.739671] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114846, 'name': ReconfigVM_Task, 'duration_secs': 0.39801} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.741045] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 3d049de4-4005-4995-a170-5cd6bbf0744c/3d049de4-4005-4995-a170-5cd6bbf0744c.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1126.741181] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-666fb852-0d96-4b56-abf9-54edcc1dd3bd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.749195] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1126.749195] env[65522]: value = "task-5114847" [ 1126.749195] env[65522]: _type = "Task" [ 1126.749195] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.758304] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114847, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.993098] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.294s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1126.993733] env[65522]: DEBUG nova.compute.manager [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1126.997033] env[65522]: DEBUG oslo_concurrency.lockutils [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.138s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1126.997261] env[65522]: DEBUG nova.objects.instance [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lazy-loading 'resources' on Instance uuid 79dd37b5-6b30-48ab-9f00-78214cbd132d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1127.028479] env[65522]: DEBUG oslo_concurrency.lockutils [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] Releasing lock "refresh_cache-95896228-9a70-4f49-bbc9-8f284a666ba9" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1127.028659] env[65522]: DEBUG nova.compute.manager [req-a8d0815a-8f90-4197-9f70-a6fea82270cd req-e9058c56-0b3a-4558-9ced-9c1f8f0ff921 service nova] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Received event network-vif-deleted-bd432f78-055d-4cd1-b502-8e058700c03a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1127.057078] env[65522]: INFO nova.compute.manager [-] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Took 1.27 seconds to deallocate network for instance. [ 1127.195301] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52aaa09e-e500-75c5-a7d6-3da0ea52af43, 'name': SearchDatastore_Task, 'duration_secs': 0.010968} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.195597] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1127.195844] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1127.196091] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.196237] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1127.196413] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1127.196690] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-00f14362-7dc4-41be-8ee2-2063c2aaf75f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.206722] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1127.206886] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1127.207645] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf2b15b1-92f8-455b-a5c8-dfe749d60d7a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.213419] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Waiting for the task: (returnval){ [ 1127.213419] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dfcdff-2298-2e9f-600b-f74dbf8d066a" [ 1127.213419] env[65522]: _type = "Task" [ 1127.213419] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.223463] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dfcdff-2298-2e9f-600b-f74dbf8d066a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.259869] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114847, 'name': Rename_Task, 'duration_secs': 0.148212} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.260182] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1127.260484] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2dfcd84-4384-4934-9372-601b52c412db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.267303] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1127.267303] env[65522]: value = "task-5114848" [ 1127.267303] env[65522]: _type = "Task" [ 1127.267303] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.275874] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114848, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.449808] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e0969b-ff3b-4ad4-9c31-5a25bddd924c tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1127.450142] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee0e064c-7b26-410c-a872-ac942f35e9e2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.458124] env[65522]: DEBUG oslo_vmware.api [None req-f0e0969b-ff3b-4ad4-9c31-5a25bddd924c tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1127.458124] env[65522]: value = "task-5114849" [ 1127.458124] env[65522]: _type = "Task" [ 1127.458124] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.467518] env[65522]: DEBUG oslo_vmware.api [None req-f0e0969b-ff3b-4ad4-9c31-5a25bddd924c tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114849, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.499081] env[65522]: DEBUG nova.compute.utils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1127.501082] env[65522]: DEBUG nova.compute.manager [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1127.501199] env[65522]: DEBUG nova.network.neutron [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1127.501479] env[65522]: WARNING neutronclient.v2_0.client [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1127.501852] env[65522]: WARNING neutronclient.v2_0.client [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1127.502751] env[65522]: WARNING openstack [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.503268] env[65522]: WARNING openstack [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.566610] env[65522]: DEBUG oslo_concurrency.lockutils [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1127.572014] env[65522]: DEBUG nova.policy [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e64fcad45994566be61b44003312b02', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fdba2dbc660145b7a98af8bd3f77a035', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1127.697852] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a199d967-290c-4f9a-96bf-a93cda9a5cfa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.708277] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae32beb-c8e2-4c4d-8454-d7b4072b7c6e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.750027] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5253a494-bcc1-46e5-bed7-7b3c5ac51a7e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.755684] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dfcdff-2298-2e9f-600b-f74dbf8d066a, 'name': SearchDatastore_Task, 'duration_secs': 0.017417} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.757073] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9eeaced8-abbf-4efd-afbd-6c63762ce2ae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.764411] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97dda304-ed47-4e1f-9d76-d65559e0a850 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.770435] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Waiting for the task: (returnval){ [ 1127.770435] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5289fb5e-0af1-fc31-38a5-1636dc6aa80a" [ 1127.770435] env[65522]: _type = "Task" [ 1127.770435] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.788719] env[65522]: DEBUG nova.compute.provider_tree [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1127.794802] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114848, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.801817] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5289fb5e-0af1-fc31-38a5-1636dc6aa80a, 'name': SearchDatastore_Task, 'duration_secs': 0.016723} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.801817] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1127.801817] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 95896228-9a70-4f49-bbc9-8f284a666ba9/95896228-9a70-4f49-bbc9-8f284a666ba9.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1127.801817] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c6edd553-780e-4440-aa7d-a9a092e56028 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.810584] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Waiting for the task: (returnval){ [ 1127.810584] env[65522]: value = "task-5114850" [ 1127.810584] env[65522]: _type = "Task" [ 1127.810584] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.821113] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114850, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.911697] env[65522]: DEBUG nova.network.neutron [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Successfully created port: 0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1127.970044] env[65522]: DEBUG oslo_vmware.api [None req-f0e0969b-ff3b-4ad4-9c31-5a25bddd924c tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114849, 'name': PowerOffVM_Task, 'duration_secs': 0.21377} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.970396] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e0969b-ff3b-4ad4-9c31-5a25bddd924c tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1127.970725] env[65522]: DEBUG nova.compute.manager [None req-f0e0969b-ff3b-4ad4-9c31-5a25bddd924c tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1127.971680] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5171bad7-2eac-4207-9f2e-6eed21f40cbb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.016789] env[65522]: DEBUG nova.compute.manager [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1128.283522] env[65522]: DEBUG oslo_vmware.api [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114848, 'name': PowerOnVM_Task, 'duration_secs': 0.541287} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.283818] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1128.284063] env[65522]: INFO nova.compute.manager [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Took 7.37 seconds to spawn the instance on the hypervisor. [ 1128.284216] env[65522]: DEBUG nova.compute.manager [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1128.285072] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f28f73ff-6d39-4c1e-a52d-a82ea93964ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.296713] env[65522]: DEBUG nova.scheduler.client.report [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1128.324248] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114850, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.484487] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f0e0969b-ff3b-4ad4-9c31-5a25bddd924c tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.062s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1128.608697] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1128.804821] env[65522]: DEBUG oslo_concurrency.lockutils [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.808s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1128.815548] env[65522]: DEBUG oslo_concurrency.lockutils [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.248s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1128.815548] env[65522]: DEBUG nova.objects.instance [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lazy-loading 'resources' on Instance uuid 2b61302b-7ce1-4e8d-b981-8d4376d76c3c {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1128.816088] env[65522]: INFO nova.compute.manager [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Took 13.31 seconds to build instance. [ 1128.829926] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114850, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579626} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.830996] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 95896228-9a70-4f49-bbc9-8f284a666ba9/95896228-9a70-4f49-bbc9-8f284a666ba9.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1128.831362] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1128.831677] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-85cb7556-87f1-46b0-884c-d82dfab35853 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.836167] env[65522]: INFO nova.scheduler.client.report [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleted allocations for instance 79dd37b5-6b30-48ab-9f00-78214cbd132d [ 1128.845321] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Waiting for the task: (returnval){ [ 1128.845321] env[65522]: value = "task-5114851" [ 1128.845321] env[65522]: _type = "Task" [ 1128.845321] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.858740] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114851, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.027289] env[65522]: DEBUG nova.compute.manager [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1129.057278] env[65522]: DEBUG nova.virt.hardware [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1129.057686] env[65522]: DEBUG nova.virt.hardware [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1129.057686] env[65522]: DEBUG nova.virt.hardware [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1129.057868] env[65522]: DEBUG nova.virt.hardware [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1129.058017] env[65522]: DEBUG nova.virt.hardware [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1129.058170] env[65522]: DEBUG nova.virt.hardware [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1129.058378] env[65522]: DEBUG nova.virt.hardware [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1129.058543] env[65522]: DEBUG nova.virt.hardware [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1129.058724] env[65522]: DEBUG nova.virt.hardware [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1129.058886] env[65522]: DEBUG nova.virt.hardware [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1129.059069] env[65522]: DEBUG nova.virt.hardware [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1129.059979] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ecc2456-c6b4-46b4-8cfa-7e882198d385 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.070018] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a217eed-6d50-4033-8b8b-85359be5218c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.324115] env[65522]: DEBUG oslo_concurrency.lockutils [None req-54c63939-0310-4a6d-b8cf-0ec334dfa096 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "3d049de4-4005-4995-a170-5cd6bbf0744c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.827s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1129.342363] env[65522]: DEBUG nova.compute.manager [req-639581e6-0619-4113-9ff1-6c23e1c439a2 req-881bd5a3-aaa9-4cca-b2fe-bbe439bcd9a5 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Received event network-vif-plugged-0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1129.342637] env[65522]: DEBUG oslo_concurrency.lockutils [req-639581e6-0619-4113-9ff1-6c23e1c439a2 req-881bd5a3-aaa9-4cca-b2fe-bbe439bcd9a5 service nova] Acquiring lock "2981b635-43c8-4bd6-9991-e6af0be82f3c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1129.342801] env[65522]: DEBUG oslo_concurrency.lockutils [req-639581e6-0619-4113-9ff1-6c23e1c439a2 req-881bd5a3-aaa9-4cca-b2fe-bbe439bcd9a5 service nova] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1129.343146] env[65522]: DEBUG oslo_concurrency.lockutils [req-639581e6-0619-4113-9ff1-6c23e1c439a2 req-881bd5a3-aaa9-4cca-b2fe-bbe439bcd9a5 service nova] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1129.343146] env[65522]: DEBUG nova.compute.manager [req-639581e6-0619-4113-9ff1-6c23e1c439a2 req-881bd5a3-aaa9-4cca-b2fe-bbe439bcd9a5 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] No waiting events found dispatching network-vif-plugged-0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1129.343387] env[65522]: WARNING nova.compute.manager [req-639581e6-0619-4113-9ff1-6c23e1c439a2 req-881bd5a3-aaa9-4cca-b2fe-bbe439bcd9a5 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Received unexpected event network-vif-plugged-0c495486-a850-4368-95a9-03fac8060924 for instance with vm_state building and task_state spawning. [ 1129.348997] env[65522]: DEBUG oslo_concurrency.lockutils [None req-01afc083-2906-454c-9b2e-fcdbb408cea8 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "79dd37b5-6b30-48ab-9f00-78214cbd132d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.206s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1129.360959] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114851, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.250163} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.361255] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1129.362075] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16b66ae-c951-486a-a37a-7d4abab45aa9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.394190] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 95896228-9a70-4f49-bbc9-8f284a666ba9/95896228-9a70-4f49-bbc9-8f284a666ba9.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1129.399345] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c82cf33-38a2-48d2-99a7-b99e1af137c0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.424401] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Waiting for the task: (returnval){ [ 1129.424401] env[65522]: value = "task-5114852" [ 1129.424401] env[65522]: _type = "Task" [ 1129.424401] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.434316] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114852, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.449165] env[65522]: DEBUG nova.network.neutron [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Successfully updated port: 0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1129.563995] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f45bd054-4476-4b6e-ad2b-e1913b645f94 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.572420] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b44cd5-c33b-4ab4-a1dc-6a0d30e6b722 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.605487] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6325b95a-5317-4e88-bebf-597fa0dab0cd tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "3d049de4-4005-4995-a170-5cd6bbf0744c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1129.605824] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6325b95a-5317-4e88-bebf-597fa0dab0cd tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "3d049de4-4005-4995-a170-5cd6bbf0744c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1129.606059] env[65522]: DEBUG nova.compute.manager [None req-6325b95a-5317-4e88-bebf-597fa0dab0cd tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1129.607324] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0d5e424-2ced-45ed-bfa2-7010dffa9333 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.611109] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c126331-83ad-43a0-9268-df4f001404ca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.622519] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40cc64f-eaf2-4326-adcb-523d21cc9d61 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.627427] env[65522]: DEBUG nova.compute.manager [None req-6325b95a-5317-4e88-bebf-597fa0dab0cd tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65522) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1129.628376] env[65522]: DEBUG nova.objects.instance [None req-6325b95a-5317-4e88-bebf-597fa0dab0cd tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lazy-loading 'flavor' on Instance uuid 3d049de4-4005-4995-a170-5cd6bbf0744c {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.643693] env[65522]: DEBUG nova.compute.provider_tree [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.681571] env[65522]: DEBUG nova.compute.manager [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Stashing vm_state: stopped {{(pid=65522) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1129.755975] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Acquiring lock "f6980679-bc2e-442f-b424-8cf80de86c02" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1129.756215] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Lock "f6980679-bc2e-442f-b424-8cf80de86c02" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1129.756456] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Acquiring lock "f6980679-bc2e-442f-b424-8cf80de86c02-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1129.756644] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Lock "f6980679-bc2e-442f-b424-8cf80de86c02-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1129.756825] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Lock "f6980679-bc2e-442f-b424-8cf80de86c02-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1129.759233] env[65522]: INFO nova.compute.manager [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Terminating instance [ 1129.935880] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114852, 'name': ReconfigVM_Task, 'duration_secs': 0.350388} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.936050] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 95896228-9a70-4f49-bbc9-8f284a666ba9/95896228-9a70-4f49-bbc9-8f284a666ba9.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1129.936691] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-090f620c-8a7d-443c-93b6-a61db77306ca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.944639] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Waiting for the task: (returnval){ [ 1129.944639] env[65522]: value = "task-5114853" [ 1129.944639] env[65522]: _type = "Task" [ 1129.944639] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.956606] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.956606] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquired lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1129.957017] env[65522]: DEBUG nova.network.neutron [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1129.962665] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114853, 'name': Rename_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.147151] env[65522]: DEBUG nova.scheduler.client.report [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1130.206905] env[65522]: DEBUG oslo_concurrency.lockutils [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1130.263812] env[65522]: DEBUG nova.compute.manager [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1130.264236] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e33b7d96-065d-4bcd-8523-abafb1683da1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.275680] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974ba90f-8553-4ac3-8051-939c0d84bb51 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.315028] env[65522]: WARNING nova.virt.vmwareapi.driver [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance f6980679-bc2e-442f-b424-8cf80de86c02 could not be found. [ 1130.315316] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1130.315813] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-578720c2-bb01-4667-bb90-84078f173e61 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.328223] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698cdd4f-3221-4f63-900c-8d2040d7923e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.367319] env[65522]: WARNING nova.virt.vmwareapi.vmops [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f6980679-bc2e-442f-b424-8cf80de86c02 could not be found. [ 1130.367319] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1130.367524] env[65522]: INFO nova.compute.manager [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Took 0.10 seconds to destroy the instance on the hypervisor. [ 1130.367767] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1130.368034] env[65522]: DEBUG nova.compute.manager [-] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1130.368136] env[65522]: DEBUG nova.network.neutron [-] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1130.368380] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1130.368890] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.369154] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.450261] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1130.459986] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114853, 'name': Rename_Task, 'duration_secs': 0.159476} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.460408] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1130.461053] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dced177b-9f1f-4721-9b69-adcf2a8194e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.466427] env[65522]: WARNING openstack [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1130.466427] env[65522]: WARNING openstack [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1130.482013] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Waiting for the task: (returnval){ [ 1130.482013] env[65522]: value = "task-5114854" [ 1130.482013] env[65522]: _type = "Task" [ 1130.482013] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.502717] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114854, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.555032] env[65522]: DEBUG nova.network.neutron [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1130.636843] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6325b95a-5317-4e88-bebf-597fa0dab0cd tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1130.637253] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c2e0964-9665-4399-b12f-fa500b04651e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.646548] env[65522]: DEBUG oslo_vmware.api [None req-6325b95a-5317-4e88-bebf-597fa0dab0cd tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1130.646548] env[65522]: value = "task-5114855" [ 1130.646548] env[65522]: _type = "Task" [ 1130.646548] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.653081] env[65522]: DEBUG oslo_concurrency.lockutils [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.839s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1130.660762] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.052s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1130.663846] env[65522]: DEBUG oslo_vmware.api [None req-6325b95a-5317-4e88-bebf-597fa0dab0cd tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114855, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.698650] env[65522]: INFO nova.scheduler.client.report [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleted allocations for instance 2b61302b-7ce1-4e8d-b981-8d4376d76c3c [ 1130.870476] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-069ccd85-dcec-4413-a6ec-c2f2a66cfe95 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.878778] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98389399-cc92-4c92-a0c6-d985ac105f82 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.913743] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae945d9e-cf7c-4ea9-adcf-10e4d90b68ce {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.923842] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cac29cf-7164-43b0-9c53-aa06c9b92cc5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.941168] env[65522]: DEBUG nova.compute.provider_tree [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1130.995917] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114854, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.065928] env[65522]: WARNING openstack [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1131.065928] env[65522]: WARNING openstack [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.164673] env[65522]: DEBUG oslo_vmware.api [None req-6325b95a-5317-4e88-bebf-597fa0dab0cd tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114855, 'name': PowerOffVM_Task, 'duration_secs': 0.243104} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.164962] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6325b95a-5317-4e88-bebf-597fa0dab0cd tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1131.165162] env[65522]: DEBUG nova.compute.manager [None req-6325b95a-5317-4e88-bebf-597fa0dab0cd tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1131.165971] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a91e8c-0bf6-412e-a586-68c0b49e9423 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.203103] env[65522]: WARNING neutronclient.v2_0.client [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1131.203487] env[65522]: WARNING openstack [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1131.203809] env[65522]: WARNING openstack [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.214825] env[65522]: DEBUG oslo_concurrency.lockutils [None req-732da285-7150-4ca7-a0d9-3d82b70d583a tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "2b61302b-7ce1-4e8d-b981-8d4376d76c3c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.127s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1131.334928] env[65522]: DEBUG nova.network.neutron [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Updating instance_info_cache with network_info: [{"id": "0c495486-a850-4368-95a9-03fac8060924", "address": "fa:16:3e:a7:35:38", "network": {"id": "2eb73cd0-541f-42e9-8774-600c5c1e9200", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-939435024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdba2dbc660145b7a98af8bd3f77a035", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c495486-a8", "ovs_interfaceid": "0c495486-a850-4368-95a9-03fac8060924", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1131.445303] env[65522]: DEBUG nova.scheduler.client.report [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1131.496949] env[65522]: DEBUG oslo_vmware.api [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114854, 'name': PowerOnVM_Task, 'duration_secs': 0.595491} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.497532] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1131.497912] env[65522]: INFO nova.compute.manager [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Took 8.81 seconds to spawn the instance on the hypervisor. [ 1131.498305] env[65522]: DEBUG nova.compute.manager [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1131.499308] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-382a0fa3-2d4a-471d-8216-ed7c5bce0c24 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.546530] env[65522]: DEBUG nova.network.neutron [-] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1131.680112] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6325b95a-5317-4e88-bebf-597fa0dab0cd tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "3d049de4-4005-4995-a170-5cd6bbf0744c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.074s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1131.726404] env[65522]: DEBUG nova.compute.manager [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Received event network-changed-0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1131.726710] env[65522]: DEBUG nova.compute.manager [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Refreshing instance network info cache due to event network-changed-0c495486-a850-4368-95a9-03fac8060924. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1131.727268] env[65522]: DEBUG oslo_concurrency.lockutils [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] Acquiring lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.839039] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Releasing lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.839126] env[65522]: DEBUG nova.compute.manager [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Instance network_info: |[{"id": "0c495486-a850-4368-95a9-03fac8060924", "address": "fa:16:3e:a7:35:38", "network": {"id": "2eb73cd0-541f-42e9-8774-600c5c1e9200", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-939435024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdba2dbc660145b7a98af8bd3f77a035", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c495486-a8", "ovs_interfaceid": "0c495486-a850-4368-95a9-03fac8060924", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1131.839453] env[65522]: DEBUG oslo_concurrency.lockutils [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] Acquired lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1131.839646] env[65522]: DEBUG nova.network.neutron [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Refreshing network info cache for port 0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1131.841091] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:35:38', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c495486-a850-4368-95a9-03fac8060924', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1131.849654] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Creating folder: Project (fdba2dbc660145b7a98af8bd3f77a035). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1131.851062] env[65522]: WARNING neutronclient.v2_0.client [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1131.851696] env[65522]: WARNING openstack [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1131.852095] env[65522]: WARNING openstack [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.859320] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-969eb6aa-c478-42e8-bf69-91031bcf3efc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.871707] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Created folder: Project (fdba2dbc660145b7a98af8bd3f77a035) in parent group-v994660. [ 1131.871868] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Creating folder: Instances. Parent ref: group-v994941. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1131.872199] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1a1bbfab-4066-4ed0-8602-d80426a83136 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.884506] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Created folder: Instances in parent group-v994941. [ 1131.885077] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1131.887331] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1131.887331] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b62466d0-0920-48d6-9e69-bb466636b0b6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.910457] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1131.910457] env[65522]: value = "task-5114858" [ 1131.910457] env[65522]: _type = "Task" [ 1131.910457] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.923085] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114858, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.951296] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.290s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1131.951714] env[65522]: INFO nova.compute.manager [None req-02af301c-a9f5-4b65-be60-a7c05699beab tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Successfully reverted task state from rebuilding on failure for instance. [ 1131.958248] env[65522]: DEBUG oslo_concurrency.lockutils [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.751s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1132.022899] env[65522]: INFO nova.compute.manager [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Took 16.49 seconds to build instance. [ 1132.049264] env[65522]: INFO nova.compute.manager [-] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Took 1.68 seconds to deallocate network for instance. [ 1132.170032] env[65522]: WARNING openstack [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.170477] env[65522]: WARNING openstack [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.279394] env[65522]: WARNING neutronclient.v2_0.client [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1132.280228] env[65522]: WARNING openstack [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.280560] env[65522]: WARNING openstack [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.386126] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "3d049de4-4005-4995-a170-5cd6bbf0744c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1132.386720] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "3d049de4-4005-4995-a170-5cd6bbf0744c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1132.386720] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "3d049de4-4005-4995-a170-5cd6bbf0744c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1132.388117] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "3d049de4-4005-4995-a170-5cd6bbf0744c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1132.388117] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "3d049de4-4005-4995-a170-5cd6bbf0744c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1132.390575] env[65522]: INFO nova.compute.manager [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Terminating instance [ 1132.421681] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114858, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.472027] env[65522]: INFO nova.compute.claims [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1132.477335] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1132.477699] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1132.481464] env[65522]: DEBUG nova.network.neutron [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Updated VIF entry in instance network info cache for port 0c495486-a850-4368-95a9-03fac8060924. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1132.482226] env[65522]: DEBUG nova.network.neutron [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Updating instance_info_cache with network_info: [{"id": "0c495486-a850-4368-95a9-03fac8060924", "address": "fa:16:3e:a7:35:38", "network": {"id": "2eb73cd0-541f-42e9-8774-600c5c1e9200", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-939435024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdba2dbc660145b7a98af8bd3f77a035", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c495486-a8", "ovs_interfaceid": "0c495486-a850-4368-95a9-03fac8060924", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1132.525116] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a1e503ec-0de5-4765-90af-4a6de05bb3b8 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Lock "95896228-9a70-4f49-bbc9-8f284a666ba9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.997s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1132.602595] env[65522]: INFO nova.compute.manager [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Took 0.55 seconds to detach 1 volumes for instance. [ 1132.605638] env[65522]: DEBUG nova.compute.manager [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Deleting volume: b4d96581-08cb-4fb8-98a2-bc1957d04206 {{(pid=65522) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3319}} [ 1132.829799] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "2eb8e092-255f-4997-8f10-a0cce0f4d503" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1132.830652] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "2eb8e092-255f-4997-8f10-a0cce0f4d503" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1132.896389] env[65522]: DEBUG nova.compute.manager [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1132.896461] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1132.898069] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bafdc6ed-acfa-4020-b7f6-691a021b20f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.906945] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1132.907253] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-896d3228-1cba-45bb-80aa-021e2e986e28 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.920065] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114858, 'name': CreateVM_Task, 'duration_secs': 0.604487} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.920180] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1132.920701] env[65522]: WARNING neutronclient.v2_0.client [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1132.921076] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.921230] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1132.921539] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1132.921849] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1aedf64-b3b1-4ae9-a60d-473a67e30ffc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.926982] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1132.926982] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52174cd6-8081-53de-b6bb-f497e402d80a" [ 1132.926982] env[65522]: _type = "Task" [ 1132.926982] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.937248] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52174cd6-8081-53de-b6bb-f497e402d80a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.979146] env[65522]: INFO nova.compute.resource_tracker [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating resource usage from migration 45fe253c-dd27-4c78-85ab-b18d25cd3f71 [ 1132.984050] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1132.984050] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1132.984231] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleting the datastore file [datastore1] 3d049de4-4005-4995-a170-5cd6bbf0744c {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1132.984545] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c060355-7cd6-4ece-b010-621ca34e2184 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.986494] env[65522]: DEBUG nova.compute.manager [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1132.989329] env[65522]: DEBUG oslo_concurrency.lockutils [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] Releasing lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1132.989558] env[65522]: DEBUG nova.compute.manager [req-3cfccaea-2029-4528-99a3-42659568f876 req-09694fc2-219e-4a0d-b865-ef4f6213cebc service nova] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Received event network-vif-deleted-01011dd2-f801-46a0-8528-8305d2c08ee9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1132.995240] env[65522]: DEBUG oslo_vmware.api [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1132.995240] env[65522]: value = "task-5114861" [ 1132.995240] env[65522]: _type = "Task" [ 1132.995240] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.007199] env[65522]: DEBUG oslo_vmware.api [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114861, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.149618] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1133.177048] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cef133c-155a-45f4-aab9-adf9a5c3dbc8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.186492] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842555bb-2ddc-4793-ae97-d4d0136411f8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.225750] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3172c69-135d-4acb-b146-443157785c6a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.235457] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f85a28d1-b541-4768-9816-ff3118deba0a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.252388] env[65522]: DEBUG nova.compute.provider_tree [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1133.334879] env[65522]: DEBUG nova.compute.manager [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1133.439381] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52174cd6-8081-53de-b6bb-f497e402d80a, 'name': SearchDatastore_Task, 'duration_secs': 0.015963} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.440365] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1133.440624] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1133.440856] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.440999] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1133.441231] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1133.441682] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b2f29532-7389-41cf-9be4-a6139c1df49c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.451245] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1133.451439] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1133.452248] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42000a5c-2c6e-47bc-982e-815f131747f9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.458429] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1133.458429] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ff0d08-1f90-f405-f5a3-161706dce34d" [ 1133.458429] env[65522]: _type = "Task" [ 1133.458429] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.472795] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ff0d08-1f90-f405-f5a3-161706dce34d, 'name': SearchDatastore_Task, 'duration_secs': 0.011666} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.474152] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69d2d2e4-3047-4130-b8ca-3b8e4df6d544 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.481652] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1133.481652] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a2d303-2de7-f4fe-e84f-3c1132e18d69" [ 1133.481652] env[65522]: _type = "Task" [ 1133.481652] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.493900] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a2d303-2de7-f4fe-e84f-3c1132e18d69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.506037] env[65522]: DEBUG oslo_vmware.api [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114861, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.375418} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.506316] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1133.506501] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1133.506679] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1133.506851] env[65522]: INFO nova.compute.manager [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1133.507123] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1133.507327] env[65522]: DEBUG nova.compute.manager [-] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1133.507442] env[65522]: DEBUG nova.network.neutron [-] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1133.507664] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.508206] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1133.508463] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1133.525394] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1133.645117] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1133.759806] env[65522]: DEBUG nova.scheduler.client.report [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1133.866598] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1133.888586] env[65522]: DEBUG nova.compute.manager [req-8273b077-188e-417c-b306-b53bbd2a119e req-3fa43811-f6e2-4dc1-b6f7-a2857046833c service nova] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Received event network-vif-deleted-301397ef-592e-4cf7-b2e2-53586abbec6d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1133.888586] env[65522]: INFO nova.compute.manager [req-8273b077-188e-417c-b306-b53bbd2a119e req-3fa43811-f6e2-4dc1-b6f7-a2857046833c service nova] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Neutron deleted interface 301397ef-592e-4cf7-b2e2-53586abbec6d; detaching it from the instance and deleting it from the info cache [ 1133.888844] env[65522]: DEBUG nova.network.neutron [req-8273b077-188e-417c-b306-b53bbd2a119e req-3fa43811-f6e2-4dc1-b6f7-a2857046833c service nova] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1133.929230] env[65522]: DEBUG oslo_concurrency.lockutils [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Acquiring lock "95896228-9a70-4f49-bbc9-8f284a666ba9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1133.929576] env[65522]: DEBUG oslo_concurrency.lockutils [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Lock "95896228-9a70-4f49-bbc9-8f284a666ba9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1133.929845] env[65522]: DEBUG oslo_concurrency.lockutils [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Acquiring lock "95896228-9a70-4f49-bbc9-8f284a666ba9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1133.930080] env[65522]: DEBUG oslo_concurrency.lockutils [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Lock "95896228-9a70-4f49-bbc9-8f284a666ba9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1133.930260] env[65522]: DEBUG oslo_concurrency.lockutils [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Lock "95896228-9a70-4f49-bbc9-8f284a666ba9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1133.932854] env[65522]: INFO nova.compute.manager [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Terminating instance [ 1133.993940] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a2d303-2de7-f4fe-e84f-3c1132e18d69, 'name': SearchDatastore_Task, 'duration_secs': 0.013582} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.994243] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1133.994525] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 2981b635-43c8-4bd6-9991-e6af0be82f3c/2981b635-43c8-4bd6-9991-e6af0be82f3c.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1133.994876] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c237247-b486-493c-b92d-8a472e427fc8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.002735] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1134.002735] env[65522]: value = "task-5114862" [ 1134.002735] env[65522]: _type = "Task" [ 1134.002735] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.012357] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114862, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.269049] env[65522]: DEBUG oslo_concurrency.lockutils [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.311s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1134.269049] env[65522]: INFO nova.compute.manager [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Migrating [ 1134.275548] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.127s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.275804] env[65522]: DEBUG nova.objects.instance [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Lazy-loading 'resources' on Instance uuid f6980679-bc2e-442f-b424-8cf80de86c02 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.359265] env[65522]: DEBUG nova.network.neutron [-] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1134.391592] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e1587d53-1fd0-4e07-8d47-27e1d68c942e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.403010] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a026811-3ed4-4c5d-81a6-a500df82a690 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.443451] env[65522]: DEBUG nova.compute.manager [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1134.443677] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1134.444073] env[65522]: DEBUG nova.compute.manager [req-8273b077-188e-417c-b306-b53bbd2a119e req-3fa43811-f6e2-4dc1-b6f7-a2857046833c service nova] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Detach interface failed, port_id=301397ef-592e-4cf7-b2e2-53586abbec6d, reason: Instance 3d049de4-4005-4995-a170-5cd6bbf0744c could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1134.445154] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a120ec-4c61-4a68-87da-2232f2fc0873 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.454037] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1134.454037] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23993b11-96f4-48af-8ba8-9ecffdf22206 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.461338] env[65522]: DEBUG oslo_vmware.api [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Waiting for the task: (returnval){ [ 1134.461338] env[65522]: value = "task-5114863" [ 1134.461338] env[65522]: _type = "Task" [ 1134.461338] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.471408] env[65522]: DEBUG oslo_vmware.api [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114863, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.515708] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114862, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.789488] env[65522]: DEBUG oslo_concurrency.lockutils [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.789488] env[65522]: DEBUG oslo_concurrency.lockutils [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1134.789488] env[65522]: DEBUG nova.network.neutron [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1134.864436] env[65522]: INFO nova.compute.manager [-] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Took 1.36 seconds to deallocate network for instance. [ 1134.974500] env[65522]: DEBUG oslo_vmware.api [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114863, 'name': PowerOffVM_Task, 'duration_secs': 0.389356} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.974785] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1134.974942] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1134.975219] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d9f69c2-8c59-443b-bcc3-bdbf25b63f30 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.979684] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f789c2ab-97c7-49b4-a68d-a0876a2ff8cc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.986986] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b7e66b-7d08-4746-a327-3f4cd7973f00 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.026047] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa851c8e-8231-46ba-9097-e4e405785eac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.036292] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e89a709-82ff-4b5c-ba63-9819321a3266 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.040352] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114862, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.846339} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.041717] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 2981b635-43c8-4bd6-9991-e6af0be82f3c/2981b635-43c8-4bd6-9991-e6af0be82f3c.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1135.041949] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1135.042641] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1135.042833] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1135.043010] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Deleting the datastore file [datastore1] 95896228-9a70-4f49-bbc9-8f284a666ba9 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1135.043660] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c1536e0-56b6-4048-8441-5e877f143238 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.045909] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9a893b68-374a-4628-a0d1-57d50f4a113c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.055725] env[65522]: DEBUG nova.compute.provider_tree [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1135.063333] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1135.063333] env[65522]: value = "task-5114866" [ 1135.063333] env[65522]: _type = "Task" [ 1135.063333] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.065631] env[65522]: DEBUG oslo_vmware.api [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Waiting for the task: (returnval){ [ 1135.065631] env[65522]: value = "task-5114865" [ 1135.065631] env[65522]: _type = "Task" [ 1135.065631] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.087101] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114866, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.092327] env[65522]: DEBUG oslo_vmware.api [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114865, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.292888] env[65522]: WARNING neutronclient.v2_0.client [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.293642] env[65522]: WARNING openstack [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.294143] env[65522]: WARNING openstack [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.375103] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1135.410369] env[65522]: WARNING openstack [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.410824] env[65522]: WARNING openstack [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.474118] env[65522]: WARNING neutronclient.v2_0.client [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.474841] env[65522]: WARNING openstack [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.475429] env[65522]: WARNING openstack [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.559178] env[65522]: DEBUG nova.scheduler.client.report [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1135.568713] env[65522]: DEBUG nova.network.neutron [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance_info_cache with network_info: [{"id": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "address": "fa:16:3e:c3:12:7d", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped6e9224-69", "ovs_interfaceid": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1135.584267] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114866, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09072} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.588939] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1135.589721] env[65522]: DEBUG oslo_vmware.api [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Task: {'id': task-5114865, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.238596} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.590473] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca865157-eb35-46a0-8c0b-cc721f7449b7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.593196] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1135.593381] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1135.593558] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1135.593724] env[65522]: INFO nova.compute.manager [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1135.593973] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1135.594153] env[65522]: DEBUG nova.compute.manager [-] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1135.594283] env[65522]: DEBUG nova.network.neutron [-] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1135.594598] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.594995] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.595259] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.623640] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 2981b635-43c8-4bd6-9991-e6af0be82f3c/2981b635-43c8-4bd6-9991-e6af0be82f3c.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1135.623640] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0961fd4-9844-4c57-8143-8d82d49ff1cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.639371] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.648863] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1135.648863] env[65522]: value = "task-5114867" [ 1135.648863] env[65522]: _type = "Task" [ 1135.648863] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.657540] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114867, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.009270] env[65522]: DEBUG nova.compute.manager [req-18d6ab9d-a918-4bca-adf2-01aa001e1abe req-e73dcbef-3de3-4d8a-8ac0-9ea6e87f1ded service nova] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Received event network-vif-deleted-672d56ee-9ded-4c3a-971d-e807052d71c8 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1136.009489] env[65522]: INFO nova.compute.manager [req-18d6ab9d-a918-4bca-adf2-01aa001e1abe req-e73dcbef-3de3-4d8a-8ac0-9ea6e87f1ded service nova] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Neutron deleted interface 672d56ee-9ded-4c3a-971d-e807052d71c8; detaching it from the instance and deleting it from the info cache [ 1136.009664] env[65522]: DEBUG nova.network.neutron [req-18d6ab9d-a918-4bca-adf2-01aa001e1abe req-e73dcbef-3de3-4d8a-8ac0-9ea6e87f1ded service nova] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1136.066218] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.791s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.071628] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.543s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1136.071628] env[65522]: INFO nova.compute.claims [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1136.075831] env[65522]: DEBUG oslo_concurrency.lockutils [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.160833] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114867, 'name': ReconfigVM_Task, 'duration_secs': 0.321229} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.161706] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 2981b635-43c8-4bd6-9991-e6af0be82f3c/2981b635-43c8-4bd6-9991-e6af0be82f3c.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1136.162372] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c0147a1b-862a-4482-a049-865e93a3e11c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.170123] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1136.170123] env[65522]: value = "task-5114868" [ 1136.170123] env[65522]: _type = "Task" [ 1136.170123] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.180106] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114868, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.345226] env[65522]: DEBUG nova.network.neutron [-] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1136.512978] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eccba3eb-1127-45b1-b742-fd86639dc487 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.526353] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b1bb37-4e97-4808-a059-920501ea7f89 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.565023] env[65522]: DEBUG nova.compute.manager [req-18d6ab9d-a918-4bca-adf2-01aa001e1abe req-e73dcbef-3de3-4d8a-8ac0-9ea6e87f1ded service nova] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Detach interface failed, port_id=672d56ee-9ded-4c3a-971d-e807052d71c8, reason: Instance 95896228-9a70-4f49-bbc9-8f284a666ba9 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1136.586877] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f1d40912-3461-4eb8-b821-53b84c53d03e tempest-ServerActionsV293TestJSON-2005045633 tempest-ServerActionsV293TestJSON-2005045633-project-member] Lock "f6980679-bc2e-442f-b424-8cf80de86c02" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.831s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.679857] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114868, 'name': Rename_Task, 'duration_secs': 0.163605} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.680169] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1136.680434] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a7c3f22-3174-4b45-9790-79e3c615b4b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.687217] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1136.687217] env[65522]: value = "task-5114869" [ 1136.687217] env[65522]: _type = "Task" [ 1136.687217] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.695612] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114869, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.845624] env[65522]: INFO nova.compute.manager [-] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Took 1.25 seconds to deallocate network for instance. [ 1136.954048] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Acquiring lock "59cc1b4f-b451-41d4-a071-b1ec0751bf38" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1136.954341] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Lock "59cc1b4f-b451-41d4-a071-b1ec0751bf38" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1137.202256] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114869, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.305636] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecbb8ab-02f9-48d6-9773-eae2bb03dded {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.314060] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae8c4b7-1332-4e5a-8a0f-955c68215799 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.346347] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50bc501-2d13-46c1-a1f1-cc86da63b27e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.352575] env[65522]: DEBUG oslo_concurrency.lockutils [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1137.356514] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbdc5c9b-f88d-4929-9c6f-e9d7495c9836 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.372194] env[65522]: DEBUG nova.compute.provider_tree [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1137.456567] env[65522]: DEBUG nova.compute.manager [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1137.595357] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc35a40-e529-4cf3-a0bb-6c99dfdddf1f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.617628] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance '9f342b89-bde2-4c35-ae42-cfe1e6973b74' progress to 0 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1137.698521] env[65522]: DEBUG oslo_vmware.api [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114869, 'name': PowerOnVM_Task, 'duration_secs': 0.588498} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.698881] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1137.699104] env[65522]: INFO nova.compute.manager [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Took 8.67 seconds to spawn the instance on the hypervisor. [ 1137.699291] env[65522]: DEBUG nova.compute.manager [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1137.700097] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa9999d2-925e-4c4b-b378-7575ebbcf164 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.876375] env[65522]: DEBUG nova.scheduler.client.report [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1137.979675] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.124842] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1138.125203] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a03ad7b-e5cd-4473-8d09-32d241a41481 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.133232] env[65522]: DEBUG oslo_vmware.api [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1138.133232] env[65522]: value = "task-5114870" [ 1138.133232] env[65522]: _type = "Task" [ 1138.133232] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.142075] env[65522]: DEBUG oslo_vmware.api [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114870, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.218084] env[65522]: INFO nova.compute.manager [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Took 13.54 seconds to build instance. [ 1138.382433] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.314s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.383211] env[65522]: DEBUG nova.compute.manager [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1138.385778] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.519s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.387136] env[65522]: INFO nova.compute.claims [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1138.644138] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1138.644343] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance '9f342b89-bde2-4c35-ae42-cfe1e6973b74' progress to 17 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1138.695687] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "99353848-2f0f-4388-9fcd-91e799342386" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.695951] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "99353848-2f0f-4388-9fcd-91e799342386" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.720618] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f80a939-8cd3-4734-afcc-8c4cb06ecae6 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.053s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.891795] env[65522]: DEBUG nova.compute.utils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1138.897594] env[65522]: DEBUG nova.compute.manager [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1138.897796] env[65522]: DEBUG nova.network.neutron [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1138.898158] env[65522]: WARNING neutronclient.v2_0.client [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1138.898493] env[65522]: WARNING neutronclient.v2_0.client [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1138.899100] env[65522]: WARNING openstack [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.899501] env[65522]: WARNING openstack [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.951339] env[65522]: DEBUG nova.policy [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a074a1efea64590b1a596a9a081d158', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1546bbdca07c443d84abea1155cfde37', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1139.151706] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1139.152044] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1139.152210] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1139.152481] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1139.152661] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1139.152817] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1139.153034] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1139.153199] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1139.153364] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1139.153527] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1139.153743] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1139.159364] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db6834fa-d206-4366-80cd-18a3787909b3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.176698] env[65522]: DEBUG oslo_vmware.api [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1139.176698] env[65522]: value = "task-5114871" [ 1139.176698] env[65522]: _type = "Task" [ 1139.176698] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.186531] env[65522]: DEBUG oslo_vmware.api [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114871, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.199509] env[65522]: INFO nova.compute.manager [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Detaching volume 3c40ed31-83a3-4a51-ace6-fae3fc2c638c [ 1139.238998] env[65522]: INFO nova.virt.block_device [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Attempting to driver detach volume 3c40ed31-83a3-4a51-ace6-fae3fc2c638c from mountpoint /dev/sdb [ 1139.239302] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1139.239549] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994920', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'name': 'volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '99353848-2f0f-4388-9fcd-91e799342386', 'attached_at': '', 'detached_at': '', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'serial': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1139.240640] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3b1007-31fb-459f-87d1-8bf9143cc509 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.273050] env[65522]: DEBUG nova.network.neutron [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Successfully created port: fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1139.276421] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb40fe7d-ea5e-4038-bc86-06a5a96f2638 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.288477] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a913b93-f712-48e2-98ea-c972f379f2e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.310931] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c9430ef-fe11-48a4-b006-e52c729f2582 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.331636] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The volume has not been displaced from its original location: [datastore1] volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c/volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1139.339150] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Reconfiguring VM instance instance-00000058 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1139.339643] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1015b0d7-26bb-4681-a1c1-f4667794ae2f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.366207] env[65522]: DEBUG oslo_vmware.api [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1139.366207] env[65522]: value = "task-5114872" [ 1139.366207] env[65522]: _type = "Task" [ 1139.366207] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.381380] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "b93f2d74-a24b-4f2c-aefd-16b093107ec4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1139.381671] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "b93f2d74-a24b-4f2c-aefd-16b093107ec4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1139.382962] env[65522]: DEBUG oslo_vmware.api [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114872, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.398934] env[65522]: DEBUG nova.compute.manager [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1139.626513] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf96a74-281c-4547-a89a-3be76784b0bd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.634718] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fecc759-d2d3-4316-9737-bc2358ceca1a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.667539] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a925ff-e788-4b04-984e-c47752867b53 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.675738] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6deb6c-2cc5-4768-a61e-2ed0992ab36f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.692480] env[65522]: DEBUG nova.compute.provider_tree [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.696584] env[65522]: DEBUG oslo_vmware.api [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114871, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.876796] env[65522]: DEBUG oslo_vmware.api [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.885420] env[65522]: DEBUG nova.compute.manager [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1140.194933] env[65522]: DEBUG oslo_vmware.api [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114871, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.198026] env[65522]: DEBUG nova.scheduler.client.report [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1140.377122] env[65522]: DEBUG oslo_vmware.api [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.406470] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1140.409930] env[65522]: DEBUG nova.compute.manager [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1140.436918] env[65522]: DEBUG nova.virt.hardware [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1140.437193] env[65522]: DEBUG nova.virt.hardware [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1140.437352] env[65522]: DEBUG nova.virt.hardware [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1140.437535] env[65522]: DEBUG nova.virt.hardware [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1140.437681] env[65522]: DEBUG nova.virt.hardware [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1140.437828] env[65522]: DEBUG nova.virt.hardware [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1140.438049] env[65522]: DEBUG nova.virt.hardware [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1140.438220] env[65522]: DEBUG nova.virt.hardware [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1140.438392] env[65522]: DEBUG nova.virt.hardware [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1140.438556] env[65522]: DEBUG nova.virt.hardware [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1140.438796] env[65522]: DEBUG nova.virt.hardware [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1140.440445] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc9585d-87e8-4b65-adec-edb71a94a401 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.449205] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-980a5a22-7068-4aa2-a515-97b71947e969 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.692341] env[65522]: DEBUG oslo_vmware.api [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114871, 'name': ReconfigVM_Task, 'duration_secs': 1.145629} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.692878] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance '9f342b89-bde2-4c35-ae42-cfe1e6973b74' progress to 33 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1140.705653] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.318s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1140.705653] env[65522]: DEBUG nova.compute.manager [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1140.706757] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.332s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1140.706961] env[65522]: DEBUG nova.objects.instance [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lazy-loading 'resources' on Instance uuid 3d049de4-4005-4995-a170-5cd6bbf0744c {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.799228] env[65522]: DEBUG nova.compute.manager [req-d8657993-0182-437c-bf1b-9dc985f1bee6 req-2e0222ad-8c37-4f74-89b1-56c2b713d917 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Received event network-vif-plugged-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1140.799448] env[65522]: DEBUG oslo_concurrency.lockutils [req-d8657993-0182-437c-bf1b-9dc985f1bee6 req-2e0222ad-8c37-4f74-89b1-56c2b713d917 service nova] Acquiring lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1140.799652] env[65522]: DEBUG oslo_concurrency.lockutils [req-d8657993-0182-437c-bf1b-9dc985f1bee6 req-2e0222ad-8c37-4f74-89b1-56c2b713d917 service nova] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1140.800382] env[65522]: DEBUG oslo_concurrency.lockutils [req-d8657993-0182-437c-bf1b-9dc985f1bee6 req-2e0222ad-8c37-4f74-89b1-56c2b713d917 service nova] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1140.800382] env[65522]: DEBUG nova.compute.manager [req-d8657993-0182-437c-bf1b-9dc985f1bee6 req-2e0222ad-8c37-4f74-89b1-56c2b713d917 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] No waiting events found dispatching network-vif-plugged-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1140.800382] env[65522]: WARNING nova.compute.manager [req-d8657993-0182-437c-bf1b-9dc985f1bee6 req-2e0222ad-8c37-4f74-89b1-56c2b713d917 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Received unexpected event network-vif-plugged-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc for instance with vm_state building and task_state spawning. [ 1140.877107] env[65522]: DEBUG oslo_vmware.api [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114872, 'name': ReconfigVM_Task, 'duration_secs': 1.377486} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.877424] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Reconfigured VM instance instance-00000058 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1140.885025] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b1b4045-c71e-4bb0-81ac-a0a92dcc501b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.897403] env[65522]: DEBUG nova.network.neutron [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Successfully updated port: fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1140.904504] env[65522]: DEBUG oslo_vmware.api [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1140.904504] env[65522]: value = "task-5114873" [ 1140.904504] env[65522]: _type = "Task" [ 1140.904504] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.914865] env[65522]: DEBUG oslo_vmware.api [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114873, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.199671] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1141.199920] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1141.200085] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1141.200291] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1141.200446] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1141.200557] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1141.200757] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1141.200913] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1141.201147] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1141.201429] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1141.201794] env[65522]: DEBUG nova.virt.hardware [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1141.208744] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Reconfiguring VM instance instance-00000053 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1141.210053] env[65522]: DEBUG nova.compute.utils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1141.211408] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec124034-d092-48c8-bbda-0f55321959c2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.225878] env[65522]: DEBUG nova.compute.manager [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1141.228478] env[65522]: DEBUG nova.compute.manager [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1141.228650] env[65522]: DEBUG nova.network.neutron [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1141.228934] env[65522]: WARNING neutronclient.v2_0.client [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1141.229257] env[65522]: WARNING neutronclient.v2_0.client [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1141.230171] env[65522]: WARNING openstack [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.230570] env[65522]: WARNING openstack [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.247665] env[65522]: DEBUG oslo_vmware.api [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1141.247665] env[65522]: value = "task-5114874" [ 1141.247665] env[65522]: _type = "Task" [ 1141.247665] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.260915] env[65522]: DEBUG oslo_vmware.api [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114874, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.282342] env[65522]: DEBUG nova.policy [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a97b2f232f7844c993dccdc386ce3b2c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9614f18e01594bf083e619c11760fe36', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1141.403647] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.404047] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1141.404318] env[65522]: DEBUG nova.network.neutron [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1141.415768] env[65522]: DEBUG oslo_vmware.api [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114873, 'name': ReconfigVM_Task, 'duration_secs': 0.241646} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.418649] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994920', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'name': 'volume-3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '99353848-2f0f-4388-9fcd-91e799342386', 'attached_at': '', 'detached_at': '', 'volume_id': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c', 'serial': '3c40ed31-83a3-4a51-ace6-fae3fc2c638c'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1141.446798] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36abbc6d-4764-48a1-892a-df06eca4d917 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.456356] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b71c7e5e-0446-4ae6-a969-47fdea74d11f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.499729] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d139f0-795a-4b25-ab3a-9cd0fdcdd08d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.518391] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4228a12-beec-44dd-bd34-f88aa8e9c07c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.556293] env[65522]: DEBUG nova.compute.provider_tree [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.661913] env[65522]: DEBUG nova.network.neutron [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Successfully created port: 35aba130-3bb2-45e1-9593-b70b82343229 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1141.758553] env[65522]: DEBUG oslo_vmware.api [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114874, 'name': ReconfigVM_Task, 'duration_secs': 0.188543} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.758849] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Reconfigured VM instance instance-00000053 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1141.759822] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a159c751-cfc5-4877-b3f3-2d106c1b24f8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.783867] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 9f342b89-bde2-4c35-ae42-cfe1e6973b74/9f342b89-bde2-4c35-ae42-cfe1e6973b74.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1141.784809] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed1a7b87-2e94-4dcf-b230-b6189823d5f8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.803933] env[65522]: DEBUG oslo_vmware.api [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1141.803933] env[65522]: value = "task-5114875" [ 1141.803933] env[65522]: _type = "Task" [ 1141.803933] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.812976] env[65522]: DEBUG oslo_vmware.api [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114875, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.907880] env[65522]: WARNING openstack [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.909123] env[65522]: WARNING openstack [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.946482] env[65522]: DEBUG nova.network.neutron [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1141.975646] env[65522]: WARNING openstack [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.976052] env[65522]: WARNING openstack [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.984177] env[65522]: DEBUG nova.objects.instance [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lazy-loading 'flavor' on Instance uuid 99353848-2f0f-4388-9fcd-91e799342386 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1142.053692] env[65522]: WARNING neutronclient.v2_0.client [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1142.054399] env[65522]: WARNING openstack [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1142.054762] env[65522]: WARNING openstack [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1142.063057] env[65522]: DEBUG nova.scheduler.client.report [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1142.143984] env[65522]: DEBUG nova.network.neutron [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Updating instance_info_cache with network_info: [{"id": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "address": "fa:16:3e:c3:dd:ad", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfed73ed8-c0", "ovs_interfaceid": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1142.248086] env[65522]: DEBUG nova.compute.manager [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1142.275381] env[65522]: DEBUG nova.virt.hardware [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1142.275750] env[65522]: DEBUG nova.virt.hardware [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1142.275958] env[65522]: DEBUG nova.virt.hardware [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1142.276318] env[65522]: DEBUG nova.virt.hardware [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1142.276458] env[65522]: DEBUG nova.virt.hardware [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1142.276602] env[65522]: DEBUG nova.virt.hardware [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1142.276871] env[65522]: DEBUG nova.virt.hardware [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1142.277096] env[65522]: DEBUG nova.virt.hardware [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1142.277474] env[65522]: DEBUG nova.virt.hardware [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1142.277510] env[65522]: DEBUG nova.virt.hardware [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1142.277736] env[65522]: DEBUG nova.virt.hardware [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1142.278941] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74aff285-4eb7-4e2c-9d42-cd6c29c589f8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.291355] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7b23bc-bdc6-4849-a865-50d1786122ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.323232] env[65522]: DEBUG oslo_vmware.api [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114875, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.568095] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.861s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1142.571926] env[65522]: DEBUG oslo_concurrency.lockutils [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.219s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1142.571926] env[65522]: DEBUG nova.objects.instance [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Lazy-loading 'resources' on Instance uuid 95896228-9a70-4f49-bbc9-8f284a666ba9 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1142.595951] env[65522]: INFO nova.scheduler.client.report [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleted allocations for instance 3d049de4-4005-4995-a170-5cd6bbf0744c [ 1142.646631] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1142.647070] env[65522]: DEBUG nova.compute.manager [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Instance network_info: |[{"id": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "address": "fa:16:3e:c3:dd:ad", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfed73ed8-c0", "ovs_interfaceid": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1142.647936] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:dd:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4c5eb94-841c-4713-985a-8fc4117fbaf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1142.656012] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1142.656754] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1142.657778] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-563db3d8-0b7d-4abd-a367-a5ef060fcaf7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.677739] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1142.677739] env[65522]: value = "task-5114876" [ 1142.677739] env[65522]: _type = "Task" [ 1142.677739] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.687341] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114876, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.822112] env[65522]: DEBUG oslo_vmware.api [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114875, 'name': ReconfigVM_Task, 'duration_secs': 0.542646} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.822424] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 9f342b89-bde2-4c35-ae42-cfe1e6973b74/9f342b89-bde2-4c35-ae42-cfe1e6973b74.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1142.822765] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance '9f342b89-bde2-4c35-ae42-cfe1e6973b74' progress to 50 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1142.842180] env[65522]: DEBUG nova.compute.manager [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Received event network-changed-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1142.842180] env[65522]: DEBUG nova.compute.manager [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Refreshing instance network info cache due to event network-changed-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1142.842180] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] Acquiring lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.842522] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] Acquired lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1142.842646] env[65522]: DEBUG nova.network.neutron [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Refreshing network info cache for port fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1142.992432] env[65522]: DEBUG oslo_concurrency.lockutils [None req-2f38796b-4a67-48f3-a262-35517b87d305 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "99353848-2f0f-4388-9fcd-91e799342386" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.295s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1143.096650] env[65522]: DEBUG nova.compute.manager [req-709dc178-d2d1-4354-a4e2-7383b7baa6e3 req-2e1c6c01-fd40-42e1-b2c4-c9b070b21097 service nova] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Received event network-vif-plugged-35aba130-3bb2-45e1-9593-b70b82343229 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1143.096761] env[65522]: DEBUG oslo_concurrency.lockutils [req-709dc178-d2d1-4354-a4e2-7383b7baa6e3 req-2e1c6c01-fd40-42e1-b2c4-c9b070b21097 service nova] Acquiring lock "2eb8e092-255f-4997-8f10-a0cce0f4d503-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1143.096977] env[65522]: DEBUG oslo_concurrency.lockutils [req-709dc178-d2d1-4354-a4e2-7383b7baa6e3 req-2e1c6c01-fd40-42e1-b2c4-c9b070b21097 service nova] Lock "2eb8e092-255f-4997-8f10-a0cce0f4d503-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1143.097163] env[65522]: DEBUG oslo_concurrency.lockutils [req-709dc178-d2d1-4354-a4e2-7383b7baa6e3 req-2e1c6c01-fd40-42e1-b2c4-c9b070b21097 service nova] Lock "2eb8e092-255f-4997-8f10-a0cce0f4d503-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1143.098184] env[65522]: DEBUG nova.compute.manager [req-709dc178-d2d1-4354-a4e2-7383b7baa6e3 req-2e1c6c01-fd40-42e1-b2c4-c9b070b21097 service nova] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] No waiting events found dispatching network-vif-plugged-35aba130-3bb2-45e1-9593-b70b82343229 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1143.098184] env[65522]: WARNING nova.compute.manager [req-709dc178-d2d1-4354-a4e2-7383b7baa6e3 req-2e1c6c01-fd40-42e1-b2c4-c9b070b21097 service nova] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Received unexpected event network-vif-plugged-35aba130-3bb2-45e1-9593-b70b82343229 for instance with vm_state building and task_state spawning. [ 1143.104262] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9c6c96f1-4b9f-4ae1-9016-89e697bd7eb5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "3d049de4-4005-4995-a170-5cd6bbf0744c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.718s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1143.192290] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114876, 'name': CreateVM_Task, 'duration_secs': 0.373584} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.193279] env[65522]: DEBUG nova.network.neutron [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Successfully updated port: 35aba130-3bb2-45e1-9593-b70b82343229 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1143.194609] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1143.195637] env[65522]: WARNING neutronclient.v2_0.client [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.195943] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.196293] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1143.196592] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1143.196983] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eebff69f-96be-4e70-9aab-ddfc462e09df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.205498] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1143.205498] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52736812-f6ec-48bf-3fe1-ac199aab9d47" [ 1143.205498] env[65522]: _type = "Task" [ 1143.205498] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.215657] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52736812-f6ec-48bf-3fe1-ac199aab9d47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.270931] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de35404e-4657-4b3e-aca9-d839ac83f70a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.279841] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9049faf9-f03c-4d85-8112-fd22b0ab4085 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.311922] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1cb1127-69ce-4a11-94e3-a3c03759bb12 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.320164] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d586ee4a-5851-477a-aa92-37cb350ced2f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.336438] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0d6175-0621-4046-a358-775ab7945c26 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.339480] env[65522]: DEBUG nova.compute.provider_tree [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.367120] env[65522]: WARNING neutronclient.v2_0.client [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.368038] env[65522]: WARNING openstack [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.368521] env[65522]: WARNING openstack [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.381357] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda2dee0-fa74-467a-a43c-daa8ef02414f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.407509] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance '9f342b89-bde2-4c35-ae42-cfe1e6973b74' progress to 67 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1143.498729] env[65522]: WARNING openstack [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.499215] env[65522]: WARNING openstack [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.582889] env[65522]: WARNING neutronclient.v2_0.client [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.583851] env[65522]: WARNING openstack [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.584310] env[65522]: WARNING openstack [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.676913] env[65522]: DEBUG nova.network.neutron [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Updated VIF entry in instance network info cache for port fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1143.677356] env[65522]: DEBUG nova.network.neutron [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Updating instance_info_cache with network_info: [{"id": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "address": "fa:16:3e:c3:dd:ad", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfed73ed8-c0", "ovs_interfaceid": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1143.696084] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "refresh_cache-2eb8e092-255f-4997-8f10-a0cce0f4d503" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.696331] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "refresh_cache-2eb8e092-255f-4997-8f10-a0cce0f4d503" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1143.696468] env[65522]: DEBUG nova.network.neutron [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1143.716799] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52736812-f6ec-48bf-3fe1-ac199aab9d47, 'name': SearchDatastore_Task, 'duration_secs': 0.018038} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.717124] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1143.717362] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1143.717618] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.717765] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1143.717940] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1143.718483] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29a2208f-1930-47cb-a6be-11e4620c1bd9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.727928] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1143.728163] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1143.728903] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03d1117b-1e06-4d8a-9531-a85bd3bbe25b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.734534] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1143.734534] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52266413-220d-1cc4-e984-0a477f382607" [ 1143.734534] env[65522]: _type = "Task" [ 1143.734534] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.742723] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52266413-220d-1cc4-e984-0a477f382607, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.843572] env[65522]: DEBUG nova.scheduler.client.report [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1143.918083] env[65522]: WARNING neutronclient.v2_0.client [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.955625] env[65522]: DEBUG nova.network.neutron [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Port ed6e9224-69f8-46fd-a71f-27d8bc6eaec3 binding to destination host cpu-1 is already ACTIVE {{(pid=65522) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 1144.053405] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "af9e93f9-3e6f-4267-b7d9-1961af38c84f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1144.053701] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "af9e93f9-3e6f-4267-b7d9-1961af38c84f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1144.079322] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "99353848-2f0f-4388-9fcd-91e799342386" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1144.079555] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "99353848-2f0f-4388-9fcd-91e799342386" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1144.079773] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "99353848-2f0f-4388-9fcd-91e799342386-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1144.079952] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "99353848-2f0f-4388-9fcd-91e799342386-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1144.080128] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "99353848-2f0f-4388-9fcd-91e799342386-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1144.082438] env[65522]: INFO nova.compute.manager [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Terminating instance [ 1144.180428] env[65522]: DEBUG oslo_concurrency.lockutils [req-0dd763f7-022e-4a58-9681-927d66f56cbf req-f721f6cb-b856-41f8-b672-533070536cfe service nova] Releasing lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1144.200207] env[65522]: WARNING openstack [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.200705] env[65522]: WARNING openstack [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.245219] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52266413-220d-1cc4-e984-0a477f382607, 'name': SearchDatastore_Task, 'duration_secs': 0.010312} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.246036] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2f49026-b13f-4f3a-8c17-b84c309243d2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.249090] env[65522]: DEBUG nova.network.neutron [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1144.254643] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1144.254643] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52beecf0-bc83-16a5-0664-8f0d6f1c22f5" [ 1144.254643] env[65522]: _type = "Task" [ 1144.254643] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.262487] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52beecf0-bc83-16a5-0664-8f0d6f1c22f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.269361] env[65522]: WARNING openstack [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.269728] env[65522]: WARNING openstack [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.338089] env[65522]: WARNING neutronclient.v2_0.client [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1144.338788] env[65522]: WARNING openstack [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.339174] env[65522]: WARNING openstack [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.349762] env[65522]: DEBUG oslo_concurrency.lockutils [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.779s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1144.352015] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.372s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1144.354133] env[65522]: INFO nova.compute.claims [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1144.369620] env[65522]: INFO nova.scheduler.client.report [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Deleted allocations for instance 95896228-9a70-4f49-bbc9-8f284a666ba9 [ 1144.433345] env[65522]: DEBUG nova.network.neutron [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Updating instance_info_cache with network_info: [{"id": "35aba130-3bb2-45e1-9593-b70b82343229", "address": "fa:16:3e:8c:bf:67", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35aba130-3b", "ovs_interfaceid": "35aba130-3bb2-45e1-9593-b70b82343229", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1144.446434] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.447160] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.447160] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.447283] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.448049] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.448049] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.448049] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.448049] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1144.448049] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.556604] env[65522]: DEBUG nova.compute.manager [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1144.586524] env[65522]: DEBUG nova.compute.manager [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1144.586842] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1144.589018] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc699dc-576c-453c-a213-ba34f86b7de9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.598030] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1144.598509] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-027b26bf-350a-463a-8601-4e509944f2a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.606716] env[65522]: DEBUG oslo_vmware.api [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1144.606716] env[65522]: value = "task-5114877" [ 1144.606716] env[65522]: _type = "Task" [ 1144.606716] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.617142] env[65522]: DEBUG oslo_vmware.api [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114877, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.765337] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52beecf0-bc83-16a5-0664-8f0d6f1c22f5, 'name': SearchDatastore_Task, 'duration_secs': 0.010717} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.765646] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1144.765919] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 3b82f9ce-ea59-430d-adb6-918cc6cc48a4/3b82f9ce-ea59-430d-adb6-918cc6cc48a4.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1144.766251] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f67b0fc-2f36-4b3e-9665-9ca18025a3e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.774641] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1144.774641] env[65522]: value = "task-5114878" [ 1144.774641] env[65522]: _type = "Task" [ 1144.774641] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.784250] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114878, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.878088] env[65522]: DEBUG oslo_concurrency.lockutils [None req-70174d60-c7b4-4499-b6ee-e364b3e17c35 tempest-ServerTagsTestJSON-206047152 tempest-ServerTagsTestJSON-206047152-project-member] Lock "95896228-9a70-4f49-bbc9-8f284a666ba9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.948s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1144.935905] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "refresh_cache-2eb8e092-255f-4997-8f10-a0cce0f4d503" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1144.936479] env[65522]: DEBUG nova.compute.manager [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Instance network_info: |[{"id": "35aba130-3bb2-45e1-9593-b70b82343229", "address": "fa:16:3e:8c:bf:67", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35aba130-3b", "ovs_interfaceid": "35aba130-3bb2-45e1-9593-b70b82343229", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1144.937112] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:bf:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10b81051-1eb1-406b-888c-4548c470c77e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '35aba130-3bb2-45e1-9593-b70b82343229', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1144.945243] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1144.945515] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1144.945824] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4f0e736-9878-4890-a89d-ceb8d111601c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.964356] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1144.983428] env[65522]: DEBUG oslo_concurrency.lockutils [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1144.983710] env[65522]: DEBUG oslo_concurrency.lockutils [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1144.983896] env[65522]: DEBUG oslo_concurrency.lockutils [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1144.984860] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1144.984860] env[65522]: value = "task-5114879" [ 1144.984860] env[65522]: _type = "Task" [ 1144.984860] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.996331] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114879, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.083797] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1145.118840] env[65522]: DEBUG oslo_vmware.api [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114877, 'name': PowerOffVM_Task, 'duration_secs': 0.326826} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.119211] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1145.119663] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1145.119821] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a14ddba6-d260-4106-a4b3-66177a53a35e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.138694] env[65522]: DEBUG nova.compute.manager [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Received event network-changed-35aba130-3bb2-45e1-9593-b70b82343229 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1145.139212] env[65522]: DEBUG nova.compute.manager [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Refreshing instance network info cache due to event network-changed-35aba130-3bb2-45e1-9593-b70b82343229. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1145.139212] env[65522]: DEBUG oslo_concurrency.lockutils [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] Acquiring lock "refresh_cache-2eb8e092-255f-4997-8f10-a0cce0f4d503" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.141795] env[65522]: DEBUG oslo_concurrency.lockutils [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] Acquired lock "refresh_cache-2eb8e092-255f-4997-8f10-a0cce0f4d503" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1145.141795] env[65522]: DEBUG nova.network.neutron [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Refreshing network info cache for port 35aba130-3bb2-45e1-9593-b70b82343229 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1145.218419] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1145.218755] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1145.218956] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleting the datastore file [datastore1] 99353848-2f0f-4388-9fcd-91e799342386 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1145.219299] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2fcd3eca-b8c6-4167-9bfd-328f853f3be6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.233177] env[65522]: DEBUG oslo_vmware.api [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1145.233177] env[65522]: value = "task-5114881" [ 1145.233177] env[65522]: _type = "Task" [ 1145.233177] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.245728] env[65522]: DEBUG oslo_vmware.api [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114881, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.290592] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114878, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.497365] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114879, 'name': CreateVM_Task, 'duration_secs': 0.415353} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.501306] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1145.502364] env[65522]: WARNING neutronclient.v2_0.client [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.502841] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.503026] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1145.503263] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1145.504016] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39689f45-a14d-4028-a903-b01d39dd3db0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.509711] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1145.509711] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520ffa73-86d9-37b7-9f93-8b75105dc9e8" [ 1145.509711] env[65522]: _type = "Task" [ 1145.509711] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.522832] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520ffa73-86d9-37b7-9f93-8b75105dc9e8, 'name': SearchDatastore_Task, 'duration_secs': 0.010549} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.523150] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.523382] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1145.523633] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.523794] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1145.523971] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1145.524260] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f4c089d-1cf9-4265-b206-8f6ce8b84d93 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.532977] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1145.533176] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1145.533916] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a459f5a-e906-458e-b3d3-f35637cd64ed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.542630] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1145.542630] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fcccae-4241-a7e7-0590-7ade9bebaa99" [ 1145.542630] env[65522]: _type = "Task" [ 1145.542630] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.551088] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fcccae-4241-a7e7-0590-7ade9bebaa99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.565652] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f430f224-7a94-4b44-9318-956dcc332508 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.573249] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4415b449-df7a-448c-b619-110414498234 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.617700] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf65344-1efc-4a59-bbfd-e7afabcdfea0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.626388] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-209727cc-bb16-49ed-ae91-056b0244abdf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.642031] env[65522]: DEBUG nova.compute.provider_tree [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1145.644083] env[65522]: WARNING neutronclient.v2_0.client [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.645213] env[65522]: WARNING openstack [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1145.645213] env[65522]: WARNING openstack [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1145.750623] env[65522]: DEBUG oslo_vmware.api [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114881, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.273361} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.750907] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1145.751139] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1145.751326] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1145.751500] env[65522]: INFO nova.compute.manager [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1145.751854] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1145.752197] env[65522]: DEBUG nova.compute.manager [-] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1145.752197] env[65522]: DEBUG nova.network.neutron [-] [instance: 99353848-2f0f-4388-9fcd-91e799342386] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1145.752401] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.752997] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1145.754091] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1145.782618] env[65522]: WARNING openstack [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1145.782815] env[65522]: WARNING openstack [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1145.796129] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114878, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.602531} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.796422] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 3b82f9ce-ea59-430d-adb6-918cc6cc48a4/3b82f9ce-ea59-430d-adb6-918cc6cc48a4.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1145.797859] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1145.797859] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fefa8a41-45e3-49d4-a882-5ec1e382b8c9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.808787] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1145.808787] env[65522]: value = "task-5114882" [ 1145.808787] env[65522]: _type = "Task" [ 1145.808787] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.824668] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114882, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.842537] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.988250] env[65522]: WARNING neutronclient.v2_0.client [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.055129] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fcccae-4241-a7e7-0590-7ade9bebaa99, 'name': SearchDatastore_Task, 'duration_secs': 0.026968} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.056014] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fe77d09-805f-4678-873a-f8d1a53bd498 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.062228] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1146.062228] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529e05e8-c5da-fb1a-8004-4cc4284bcde1" [ 1146.062228] env[65522]: _type = "Task" [ 1146.062228] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.071623] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529e05e8-c5da-fb1a-8004-4cc4284bcde1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.144026] env[65522]: WARNING neutronclient.v2_0.client [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.144026] env[65522]: WARNING openstack [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.144026] env[65522]: WARNING openstack [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1146.154608] env[65522]: DEBUG nova.scheduler.client.report [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1146.278246] env[65522]: DEBUG oslo_concurrency.lockutils [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.278470] env[65522]: DEBUG oslo_concurrency.lockutils [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1146.278653] env[65522]: DEBUG nova.network.neutron [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1146.323767] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114882, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.168005} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.324100] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1146.324913] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04f6f44-b722-4eb7-a4f3-099333422ae5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.349253] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 3b82f9ce-ea59-430d-adb6-918cc6cc48a4/3b82f9ce-ea59-430d-adb6-918cc6cc48a4.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1146.351356] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f824e164-aeef-448a-9d06-e6b7e0b3008b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.379540] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1146.379540] env[65522]: value = "task-5114883" [ 1146.379540] env[65522]: _type = "Task" [ 1146.379540] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.391099] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114883, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.396955] env[65522]: DEBUG nova.network.neutron [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Updated VIF entry in instance network info cache for port 35aba130-3bb2-45e1-9593-b70b82343229. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1146.398232] env[65522]: DEBUG nova.network.neutron [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Updating instance_info_cache with network_info: [{"id": "35aba130-3bb2-45e1-9593-b70b82343229", "address": "fa:16:3e:8c:bf:67", "network": {"id": "5bf2795f-ff67-4949-a804-595780b98c88", "bridge": "br-int", "label": "tempest-ServersTestJSON-2103480464-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9614f18e01594bf083e619c11760fe36", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35aba130-3b", "ovs_interfaceid": "35aba130-3bb2-45e1-9593-b70b82343229", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1146.983722] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.632s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1146.984499] env[65522]: DEBUG nova.compute.manager [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1146.992230] env[65522]: WARNING neutronclient.v2_0.client [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.992352] env[65522]: WARNING openstack [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.996106] env[65522]: WARNING openstack [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.009829] env[65522]: DEBUG nova.network.neutron [-] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1147.012077] env[65522]: DEBUG oslo_concurrency.lockutils [req-1b4e0509-ea73-44b3-86cd-8df9668e56de req-c616b53a-4e20-41cf-b7a1-027e1b918957 service nova] Releasing lock "refresh_cache-2eb8e092-255f-4997-8f10-a0cce0f4d503" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1147.027461] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.615s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1147.027461] env[65522]: INFO nova.compute.claims [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1147.034257] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529e05e8-c5da-fb1a-8004-4cc4284bcde1, 'name': SearchDatastore_Task, 'duration_secs': 0.022277} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.037209] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1147.037209] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 2eb8e092-255f-4997-8f10-a0cce0f4d503/2eb8e092-255f-4997-8f10-a0cce0f4d503.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1147.037659] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f67e646-8a4f-429c-bcad-6ee21b3b65fd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.048351] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114883, 'name': ReconfigVM_Task, 'duration_secs': 0.295774} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.050099] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 3b82f9ce-ea59-430d-adb6-918cc6cc48a4/3b82f9ce-ea59-430d-adb6-918cc6cc48a4.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1147.051077] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-052a5907-a325-4397-ba74-735abed23013 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.057603] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1147.057603] env[65522]: value = "task-5114884" [ 1147.057603] env[65522]: _type = "Task" [ 1147.057603] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.065507] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1147.065507] env[65522]: value = "task-5114885" [ 1147.065507] env[65522]: _type = "Task" [ 1147.065507] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.074705] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114884, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.082236] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114885, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.085851] env[65522]: DEBUG nova.compute.manager [req-52b68ce4-220c-42b5-8262-3289712b2785 req-4fdf9400-48c0-4b0f-9ed5-fca97a229e38 service nova] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Received event network-vif-deleted-18edb16a-a08b-4e97-9b0a-a138798ed559 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1147.180549] env[65522]: WARNING openstack [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1147.181190] env[65522]: WARNING openstack [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.263468] env[65522]: WARNING neutronclient.v2_0.client [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1147.264176] env[65522]: WARNING openstack [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1147.264549] env[65522]: WARNING openstack [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.370407] env[65522]: DEBUG nova.network.neutron [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance_info_cache with network_info: [{"id": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "address": "fa:16:3e:c3:12:7d", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped6e9224-69", "ovs_interfaceid": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1147.515448] env[65522]: DEBUG nova.compute.utils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1147.516959] env[65522]: INFO nova.compute.manager [-] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Took 1.76 seconds to deallocate network for instance. [ 1147.517479] env[65522]: DEBUG nova.compute.manager [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1147.517733] env[65522]: DEBUG nova.network.neutron [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1147.518039] env[65522]: WARNING neutronclient.v2_0.client [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1147.518700] env[65522]: WARNING neutronclient.v2_0.client [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1147.519056] env[65522]: WARNING openstack [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1147.519429] env[65522]: WARNING openstack [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.575575] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114884, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.583819] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114885, 'name': Rename_Task, 'duration_secs': 0.145804} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.587416] env[65522]: DEBUG nova.policy [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '006911f1785b4cc980466f92d8deb70d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '367e450334d147ba9537d8c8d0be7f5d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1147.589323] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1147.590122] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24abe906-f8f9-44ca-b826-d25b4dcf1021 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.598178] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1147.598178] env[65522]: value = "task-5114886" [ 1147.598178] env[65522]: _type = "Task" [ 1147.598178] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.611247] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114886, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.847598] env[65522]: DEBUG nova.compute.manager [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Stashing vm_state: active {{(pid=65522) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1147.873547] env[65522]: DEBUG oslo_concurrency.lockutils [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1147.905121] env[65522]: DEBUG nova.network.neutron [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Successfully created port: ca385019-1696-4857-b5cc-bc42db6d70f4 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1148.021750] env[65522]: DEBUG nova.compute.manager [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1148.036280] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1148.073869] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114884, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.111901] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114886, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.285097] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cadec7a-5534-4271-9761-0f50995c7a3c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.293649] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a3a753-7693-4a36-89cb-674324d66866 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.328121] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4073a4-99e4-4b6e-9887-70bd38aefc4f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.336821] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c2f5f9-a1ec-4c4e-80fa-6b8a044fee95 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.356858] env[65522]: DEBUG nova.compute.provider_tree [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1148.373225] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1148.397570] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42841012-49e5-4838-b348-83962dedccc1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.418826] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f28d75-42e4-4cc1-9333-bfebe07398f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.427190] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance '9f342b89-bde2-4c35-ae42-cfe1e6973b74' progress to 83 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1148.443316] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquiring lock "6d856111-81bd-4e2f-b7fc-e169c3e974a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1148.443686] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Lock "6d856111-81bd-4e2f-b7fc-e169c3e974a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1148.570471] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114884, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.609275] env[65522]: DEBUG oslo_vmware.api [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114886, 'name': PowerOnVM_Task, 'duration_secs': 0.630466} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.610024] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1148.610024] env[65522]: INFO nova.compute.manager [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Took 8.20 seconds to spawn the instance on the hypervisor. [ 1148.610024] env[65522]: DEBUG nova.compute.manager [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1148.611497] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd1af23-dc5c-478d-a548-30c8c89e242e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.860939] env[65522]: DEBUG nova.scheduler.client.report [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1148.934524] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-394e42da-e0d1-426a-b2e6-de7a1ac5e93a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance '9f342b89-bde2-4c35-ae42-cfe1e6973b74' progress to 100 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1148.947612] env[65522]: DEBUG nova.compute.manager [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1149.031990] env[65522]: DEBUG nova.compute.manager [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1149.058474] env[65522]: DEBUG nova.virt.hardware [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1149.058732] env[65522]: DEBUG nova.virt.hardware [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1149.058927] env[65522]: DEBUG nova.virt.hardware [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1149.059171] env[65522]: DEBUG nova.virt.hardware [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1149.059337] env[65522]: DEBUG nova.virt.hardware [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1149.059484] env[65522]: DEBUG nova.virt.hardware [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1149.059689] env[65522]: DEBUG nova.virt.hardware [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1149.059859] env[65522]: DEBUG nova.virt.hardware [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1149.060046] env[65522]: DEBUG nova.virt.hardware [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1149.060212] env[65522]: DEBUG nova.virt.hardware [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1149.060378] env[65522]: DEBUG nova.virt.hardware [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1149.061361] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d5ef2f-ea02-4876-ad30-159374d480cd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.076348] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60efa40c-aebb-4e95-84c9-9c34cd34fb9e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.080526] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114884, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.564682} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.080894] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 2eb8e092-255f-4997-8f10-a0cce0f4d503/2eb8e092-255f-4997-8f10-a0cce0f4d503.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1149.081045] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1149.081626] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c15512d-2771-43f1-821a-d5cc046c5f25 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.093664] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1149.093664] env[65522]: value = "task-5114887" [ 1149.093664] env[65522]: _type = "Task" [ 1149.093664] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.102783] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114887, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.129437] env[65522]: INFO nova.compute.manager [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Took 15.64 seconds to build instance. [ 1149.366720] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.346s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.367274] env[65522]: DEBUG nova.compute.manager [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1149.370096] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.406s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.370261] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.370412] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1149.370701] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.287s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.372108] env[65522]: INFO nova.compute.claims [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1149.375363] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2389bf60-47a5-46c7-a534-202cd1af9848 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.387233] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f1e24f-d8a9-40e3-b52c-8c3906bbd9c3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.403201] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b669ea-9c3e-465d-be78-cce7fa9ffa08 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.412062] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ec944a-607f-428a-bd5c-c2308b021e9a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.449584] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178537MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1149.449798] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.470312] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.604178] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114887, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077295} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.604469] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1149.605328] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6093097e-a746-4663-9af7-a433fc7f052e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.629588] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 2eb8e092-255f-4997-8f10-a0cce0f4d503/2eb8e092-255f-4997-8f10-a0cce0f4d503.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1149.629960] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a51b3b1-7032-48d6-bfc6-77dccbeb3c84 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.645047] env[65522]: DEBUG oslo_concurrency.lockutils [None req-39c78868-1eb4-4546-b1c3-a54e94787f95 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.167s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.652013] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1149.652013] env[65522]: value = "task-5114888" [ 1149.652013] env[65522]: _type = "Task" [ 1149.652013] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.660863] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114888, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.876697] env[65522]: DEBUG nova.compute.utils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1149.878278] env[65522]: DEBUG nova.compute.manager [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1149.878476] env[65522]: DEBUG nova.network.neutron [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1149.878795] env[65522]: WARNING neutronclient.v2_0.client [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1149.879207] env[65522]: WARNING neutronclient.v2_0.client [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1149.879961] env[65522]: WARNING openstack [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1149.880388] env[65522]: WARNING openstack [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1149.932157] env[65522]: DEBUG nova.network.neutron [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Successfully updated port: ca385019-1696-4857-b5cc-bc42db6d70f4 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1149.980567] env[65522]: DEBUG nova.policy [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e64fcad45994566be61b44003312b02', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fdba2dbc660145b7a98af8bd3f77a035', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1150.168254] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114888, 'name': ReconfigVM_Task, 'duration_secs': 0.29253} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.169873] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 2eb8e092-255f-4997-8f10-a0cce0f4d503/2eb8e092-255f-4997-8f10-a0cce0f4d503.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.171658] env[65522]: DEBUG nova.compute.manager [req-316a8b86-6a5a-4a40-a4b3-5a3263bc27d3 req-01f5247b-8b68-4a99-80fe-d54cbb4d94e3 service nova] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Received event network-vif-plugged-ca385019-1696-4857-b5cc-bc42db6d70f4 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1150.171893] env[65522]: DEBUG oslo_concurrency.lockutils [req-316a8b86-6a5a-4a40-a4b3-5a3263bc27d3 req-01f5247b-8b68-4a99-80fe-d54cbb4d94e3 service nova] Acquiring lock "59cc1b4f-b451-41d4-a071-b1ec0751bf38-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1150.172149] env[65522]: DEBUG oslo_concurrency.lockutils [req-316a8b86-6a5a-4a40-a4b3-5a3263bc27d3 req-01f5247b-8b68-4a99-80fe-d54cbb4d94e3 service nova] Lock "59cc1b4f-b451-41d4-a071-b1ec0751bf38-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1150.172356] env[65522]: DEBUG oslo_concurrency.lockutils [req-316a8b86-6a5a-4a40-a4b3-5a3263bc27d3 req-01f5247b-8b68-4a99-80fe-d54cbb4d94e3 service nova] Lock "59cc1b4f-b451-41d4-a071-b1ec0751bf38-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1150.172552] env[65522]: DEBUG nova.compute.manager [req-316a8b86-6a5a-4a40-a4b3-5a3263bc27d3 req-01f5247b-8b68-4a99-80fe-d54cbb4d94e3 service nova] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] No waiting events found dispatching network-vif-plugged-ca385019-1696-4857-b5cc-bc42db6d70f4 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1150.172820] env[65522]: WARNING nova.compute.manager [req-316a8b86-6a5a-4a40-a4b3-5a3263bc27d3 req-01f5247b-8b68-4a99-80fe-d54cbb4d94e3 service nova] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Received unexpected event network-vif-plugged-ca385019-1696-4857-b5cc-bc42db6d70f4 for instance with vm_state building and task_state spawning. [ 1150.173500] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-db4f1182-9189-459d-acde-f03e70500e73 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.181692] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1150.181692] env[65522]: value = "task-5114889" [ 1150.181692] env[65522]: _type = "Task" [ 1150.181692] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.192147] env[65522]: DEBUG nova.compute.manager [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Received event network-changed-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1150.192147] env[65522]: DEBUG nova.compute.manager [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Refreshing instance network info cache due to event network-changed-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1150.192332] env[65522]: DEBUG oslo_concurrency.lockutils [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] Acquiring lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.192484] env[65522]: DEBUG oslo_concurrency.lockutils [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] Acquired lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1150.192679] env[65522]: DEBUG nova.network.neutron [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Refreshing network info cache for port fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1150.197627] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114889, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.283154] env[65522]: DEBUG nova.network.neutron [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Successfully created port: eb11dd2a-68ae-4ed3-a4dc-51897ede63f7 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1150.391035] env[65522]: DEBUG nova.compute.manager [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1150.436834] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Acquiring lock "refresh_cache-59cc1b4f-b451-41d4-a071-b1ec0751bf38" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.437066] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Acquired lock "refresh_cache-59cc1b4f-b451-41d4-a071-b1ec0751bf38" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1150.437209] env[65522]: DEBUG nova.network.neutron [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1150.654249] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-271d184d-6177-4a1a-a4e7-12fe112f9de9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.662096] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b07a5652-08e7-4061-8592-214d47e57eb6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.695380] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ede7db5-d476-4a8c-87c3-00b2a42ebca3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.698644] env[65522]: WARNING neutronclient.v2_0.client [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1150.699300] env[65522]: WARNING openstack [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.699658] env[65522]: WARNING openstack [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.714011] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114889, 'name': Rename_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.715015] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-288d3c39-8958-4443-ad89-b4432db5e78c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.731245] env[65522]: DEBUG nova.compute.provider_tree [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1150.825304] env[65522]: WARNING openstack [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.825833] env[65522]: WARNING openstack [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.885587] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1150.885871] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1150.886067] env[65522]: DEBUG nova.compute.manager [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Going to confirm migration 6 {{(pid=65522) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1150.891126] env[65522]: WARNING neutronclient.v2_0.client [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1150.891979] env[65522]: WARNING openstack [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.892376] env[65522]: WARNING openstack [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.939997] env[65522]: WARNING openstack [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.940431] env[65522]: WARNING openstack [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.994970] env[65522]: DEBUG nova.network.neutron [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1151.033539] env[65522]: DEBUG nova.network.neutron [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Updated VIF entry in instance network info cache for port fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1151.033974] env[65522]: DEBUG nova.network.neutron [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Updating instance_info_cache with network_info: [{"id": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "address": "fa:16:3e:c3:dd:ad", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfed73ed8-c0", "ovs_interfaceid": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1151.051100] env[65522]: WARNING openstack [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.051559] env[65522]: WARNING openstack [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.189745] env[65522]: WARNING neutronclient.v2_0.client [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.190747] env[65522]: WARNING openstack [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.191903] env[65522]: WARNING openstack [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.210742] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114889, 'name': Rename_Task, 'duration_secs': 0.901151} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.211063] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1151.211339] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ecf6ba27-c0ee-455d-8893-b429a84d3ec0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.220160] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1151.220160] env[65522]: value = "task-5114890" [ 1151.220160] env[65522]: _type = "Task" [ 1151.220160] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.229425] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114890, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.234817] env[65522]: DEBUG nova.scheduler.client.report [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1151.351660] env[65522]: DEBUG nova.network.neutron [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Updating instance_info_cache with network_info: [{"id": "ca385019-1696-4857-b5cc-bc42db6d70f4", "address": "fa:16:3e:9d:51:c7", "network": {"id": "fc6b07bf-fc09-4ef6-b8c2-a83ecd6339a9", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-9814585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "367e450334d147ba9537d8c8d0be7f5d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca385019-16", "ovs_interfaceid": "ca385019-1696-4857-b5cc-bc42db6d70f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1151.392148] env[65522]: WARNING neutronclient.v2_0.client [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.406092] env[65522]: DEBUG nova.compute.manager [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1151.428573] env[65522]: WARNING neutronclient.v2_0.client [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.428962] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.429185] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1151.429401] env[65522]: DEBUG nova.network.neutron [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1151.429637] env[65522]: DEBUG nova.objects.instance [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'info_cache' on Instance uuid 9f342b89-bde2-4c35-ae42-cfe1e6973b74 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1151.434041] env[65522]: DEBUG nova.virt.hardware [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1151.434273] env[65522]: DEBUG nova.virt.hardware [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1151.434426] env[65522]: DEBUG nova.virt.hardware [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1151.434603] env[65522]: DEBUG nova.virt.hardware [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1151.434779] env[65522]: DEBUG nova.virt.hardware [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1151.434891] env[65522]: DEBUG nova.virt.hardware [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1151.435155] env[65522]: DEBUG nova.virt.hardware [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1151.435325] env[65522]: DEBUG nova.virt.hardware [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1151.435491] env[65522]: DEBUG nova.virt.hardware [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1151.435651] env[65522]: DEBUG nova.virt.hardware [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1151.435820] env[65522]: DEBUG nova.virt.hardware [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1151.436995] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7c3a29-8ac7-4638-9ff5-2e81f87fb5d0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.447777] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-950e50d5-dbd8-42eb-8f6d-94463217ca9d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.540016] env[65522]: DEBUG oslo_concurrency.lockutils [req-f938072a-0021-4793-86b3-4c3598f959e4 req-e6585d70-a01d-440b-b448-21c9af0984bd service nova] Releasing lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1151.732181] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114890, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.740638] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.370s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.741160] env[65522]: DEBUG nova.compute.manager [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1151.744177] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.710s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1151.744784] env[65522]: DEBUG nova.objects.instance [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lazy-loading 'resources' on Instance uuid 99353848-2f0f-4388-9fcd-91e799342386 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1151.854659] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Releasing lock "refresh_cache-59cc1b4f-b451-41d4-a071-b1ec0751bf38" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1151.855082] env[65522]: DEBUG nova.compute.manager [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Instance network_info: |[{"id": "ca385019-1696-4857-b5cc-bc42db6d70f4", "address": "fa:16:3e:9d:51:c7", "network": {"id": "fc6b07bf-fc09-4ef6-b8c2-a83ecd6339a9", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-9814585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "367e450334d147ba9537d8c8d0be7f5d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca385019-16", "ovs_interfaceid": "ca385019-1696-4857-b5cc-bc42db6d70f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1151.855581] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:51:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9c621a9c-66f5-426a-8aab-bd8b2e912106', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ca385019-1696-4857-b5cc-bc42db6d70f4', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1151.863773] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Creating folder: Project (367e450334d147ba9537d8c8d0be7f5d). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1151.865066] env[65522]: DEBUG nova.network.neutron [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Successfully updated port: eb11dd2a-68ae-4ed3-a4dc-51897ede63f7 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1151.866253] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ecf67a9f-3ca0-4096-b1f0-f69e4b0420a7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.878022] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Created folder: Project (367e450334d147ba9537d8c8d0be7f5d) in parent group-v994660. [ 1151.878204] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Creating folder: Instances. Parent ref: group-v994946. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1151.878461] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6fe7809b-3875-474b-b9bb-f96b531e1b50 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.889067] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Created folder: Instances in parent group-v994946. [ 1151.889338] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1151.889535] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1151.889764] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-03091744-0f84-459c-bc9b-0468220071fd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.909778] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1151.909778] env[65522]: value = "task-5114893" [ 1151.909778] env[65522]: _type = "Task" [ 1151.909778] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.918700] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114893, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.198260] env[65522]: DEBUG nova.compute.manager [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Received event network-changed-ca385019-1696-4857-b5cc-bc42db6d70f4 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1152.198260] env[65522]: DEBUG nova.compute.manager [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Refreshing instance network info cache due to event network-changed-ca385019-1696-4857-b5cc-bc42db6d70f4. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1152.198260] env[65522]: DEBUG oslo_concurrency.lockutils [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] Acquiring lock "refresh_cache-59cc1b4f-b451-41d4-a071-b1ec0751bf38" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.198260] env[65522]: DEBUG oslo_concurrency.lockutils [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] Acquired lock "refresh_cache-59cc1b4f-b451-41d4-a071-b1ec0751bf38" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1152.198706] env[65522]: DEBUG nova.network.neutron [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Refreshing network info cache for port ca385019-1696-4857-b5cc-bc42db6d70f4 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1152.231015] env[65522]: DEBUG oslo_vmware.api [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114890, 'name': PowerOnVM_Task, 'duration_secs': 0.532173} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.232408] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1152.232627] env[65522]: INFO nova.compute.manager [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Took 9.98 seconds to spawn the instance on the hypervisor. [ 1152.232851] env[65522]: DEBUG nova.compute.manager [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1152.233983] env[65522]: DEBUG nova.compute.manager [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Received event network-vif-plugged-eb11dd2a-68ae-4ed3-a4dc-51897ede63f7 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1152.234285] env[65522]: DEBUG oslo_concurrency.lockutils [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] Acquiring lock "b93f2d74-a24b-4f2c-aefd-16b093107ec4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1152.234550] env[65522]: DEBUG oslo_concurrency.lockutils [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] Lock "b93f2d74-a24b-4f2c-aefd-16b093107ec4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1152.234743] env[65522]: DEBUG oslo_concurrency.lockutils [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] Lock "b93f2d74-a24b-4f2c-aefd-16b093107ec4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1152.234893] env[65522]: DEBUG nova.compute.manager [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] No waiting events found dispatching network-vif-plugged-eb11dd2a-68ae-4ed3-a4dc-51897ede63f7 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1152.235093] env[65522]: WARNING nova.compute.manager [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Received unexpected event network-vif-plugged-eb11dd2a-68ae-4ed3-a4dc-51897ede63f7 for instance with vm_state building and task_state spawning. [ 1152.235329] env[65522]: DEBUG nova.compute.manager [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Received event network-changed-eb11dd2a-68ae-4ed3-a4dc-51897ede63f7 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1152.235495] env[65522]: DEBUG nova.compute.manager [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Refreshing instance network info cache due to event network-changed-eb11dd2a-68ae-4ed3-a4dc-51897ede63f7. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1152.235680] env[65522]: DEBUG oslo_concurrency.lockutils [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] Acquiring lock "refresh_cache-b93f2d74-a24b-4f2c-aefd-16b093107ec4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.235832] env[65522]: DEBUG oslo_concurrency.lockutils [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] Acquired lock "refresh_cache-b93f2d74-a24b-4f2c-aefd-16b093107ec4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1152.236043] env[65522]: DEBUG nova.network.neutron [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Refreshing network info cache for port eb11dd2a-68ae-4ed3-a4dc-51897ede63f7 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1152.237874] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d04c9ef-13f0-4d78-9312-b98dc4f8a283 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.249263] env[65522]: DEBUG nova.compute.utils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1152.256603] env[65522]: DEBUG nova.compute.manager [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1152.256811] env[65522]: DEBUG nova.network.neutron [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1152.257182] env[65522]: WARNING neutronclient.v2_0.client [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.257504] env[65522]: WARNING neutronclient.v2_0.client [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.258103] env[65522]: WARNING openstack [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.258538] env[65522]: WARNING openstack [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.323062] env[65522]: DEBUG nova.policy [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8f8fa586d8c74bb9805a1148fa7ed037', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f93c790ff61543bd8e134bcf9cb20bb2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1152.370065] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "refresh_cache-b93f2d74-a24b-4f2c-aefd-16b093107ec4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.422501] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114893, 'name': CreateVM_Task, 'duration_secs': 0.454558} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.422717] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1152.423226] env[65522]: WARNING neutronclient.v2_0.client [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.423586] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.423739] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1152.425070] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1152.425224] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a666789c-5014-446b-b84a-771516a67f27 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.430681] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Waiting for the task: (returnval){ [ 1152.430681] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525fc7f0-a8f9-c09e-d716-d1012d7b8d2f" [ 1152.430681] env[65522]: _type = "Task" [ 1152.430681] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.442261] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525fc7f0-a8f9-c09e-d716-d1012d7b8d2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.443944] env[65522]: WARNING neutronclient.v2_0.client [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.444596] env[65522]: WARNING openstack [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.444958] env[65522]: WARNING openstack [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.492533] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8455eb1-a0c7-4104-829e-99193ba6f622 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.501337] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18db5286-1120-4583-85a2-28ab784dc391 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.547315] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11b7898-78f2-4f76-b34c-4410042d8c4b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.557385] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904b0c89-35a8-433b-803d-2275536b10d4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.575359] env[65522]: DEBUG nova.compute.provider_tree [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1152.643077] env[65522]: DEBUG nova.network.neutron [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Successfully created port: 2ff02808-9c40-4796-95c9-84dea4f51ab9 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1152.683346] env[65522]: WARNING openstack [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.684100] env[65522]: WARNING openstack [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.703493] env[65522]: WARNING neutronclient.v2_0.client [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.703493] env[65522]: WARNING openstack [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.703493] env[65522]: WARNING openstack [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.743098] env[65522]: WARNING neutronclient.v2_0.client [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.743956] env[65522]: WARNING openstack [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.744415] env[65522]: WARNING openstack [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.756643] env[65522]: DEBUG nova.compute.manager [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1152.775418] env[65522]: INFO nova.compute.manager [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Took 18.93 seconds to build instance. [ 1152.837255] env[65522]: DEBUG nova.network.neutron [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1152.865084] env[65522]: WARNING neutronclient.v2_0.client [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1152.865816] env[65522]: WARNING openstack [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1152.866209] env[65522]: WARNING openstack [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.942605] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525fc7f0-a8f9-c09e-d716-d1012d7b8d2f, 'name': SearchDatastore_Task, 'duration_secs': 0.012046} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.942996] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1152.943267] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1152.943558] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.943728] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1152.943925] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1152.944241] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ac29c33-486c-4408-9d7e-d555818d3219 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.953979] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1152.954204] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1152.955121] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d94ef627-301f-433a-a43e-55a574980313 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.963549] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Waiting for the task: (returnval){ [ 1152.963549] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520afff7-c989-ec2e-16dd-50533e7fd286" [ 1152.963549] env[65522]: _type = "Task" [ 1152.963549] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.972390] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520afff7-c989-ec2e-16dd-50533e7fd286, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.008267] env[65522]: WARNING openstack [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1153.008681] env[65522]: WARNING openstack [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1153.046693] env[65522]: DEBUG nova.network.neutron [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1153.080400] env[65522]: DEBUG nova.network.neutron [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance_info_cache with network_info: [{"id": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "address": "fa:16:3e:c3:12:7d", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped6e9224-69", "ovs_interfaceid": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1153.082258] env[65522]: DEBUG nova.scheduler.client.report [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1153.131912] env[65522]: WARNING neutronclient.v2_0.client [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1153.133211] env[65522]: WARNING openstack [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1153.133584] env[65522]: WARNING openstack [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1153.279831] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f037e0b0-b6cf-401c-8d75-6478d97a895c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "2eb8e092-255f-4997-8f10-a0cce0f4d503" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.449s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.297736] env[65522]: DEBUG nova.network.neutron [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Updated VIF entry in instance network info cache for port ca385019-1696-4857-b5cc-bc42db6d70f4. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1153.298137] env[65522]: DEBUG nova.network.neutron [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Updating instance_info_cache with network_info: [{"id": "ca385019-1696-4857-b5cc-bc42db6d70f4", "address": "fa:16:3e:9d:51:c7", "network": {"id": "fc6b07bf-fc09-4ef6-b8c2-a83ecd6339a9", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-9814585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "367e450334d147ba9537d8c8d0be7f5d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca385019-16", "ovs_interfaceid": "ca385019-1696-4857-b5cc-bc42db6d70f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1153.474601] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520afff7-c989-ec2e-16dd-50533e7fd286, 'name': SearchDatastore_Task, 'duration_secs': 0.012218} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.475514] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06b1fe21-7650-4a19-8913-22c9a79d2249 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.481789] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Waiting for the task: (returnval){ [ 1153.481789] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5267214c-2e4d-59ec-4c41-c4d9ffd7023b" [ 1153.481789] env[65522]: _type = "Task" [ 1153.481789] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.490744] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5267214c-2e4d-59ec-4c41-c4d9ffd7023b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.549945] env[65522]: DEBUG oslo_concurrency.lockutils [req-8ca9eec6-d644-4562-adfd-07122870756a req-baa80c18-932e-4c5a-b081-2c42fb32814d service nova] Releasing lock "refresh_cache-b93f2d74-a24b-4f2c-aefd-16b093107ec4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1153.550357] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquired lock "refresh_cache-b93f2d74-a24b-4f2c-aefd-16b093107ec4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1153.550554] env[65522]: DEBUG nova.network.neutron [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1153.587647] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.843s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.590978] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1153.590978] env[65522]: DEBUG nova.objects.instance [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'migration_context' on Instance uuid 9f342b89-bde2-4c35-ae42-cfe1e6973b74 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1153.592173] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.219s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.614846] env[65522]: INFO nova.scheduler.client.report [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleted allocations for instance 99353848-2f0f-4388-9fcd-91e799342386 [ 1153.647638] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.647949] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.648224] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.648445] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.649289] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.651816] env[65522]: INFO nova.compute.manager [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Terminating instance [ 1153.768064] env[65522]: DEBUG nova.compute.manager [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1153.794408] env[65522]: DEBUG nova.virt.hardware [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1153.794665] env[65522]: DEBUG nova.virt.hardware [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1153.794818] env[65522]: DEBUG nova.virt.hardware [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1153.795009] env[65522]: DEBUG nova.virt.hardware [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1153.795188] env[65522]: DEBUG nova.virt.hardware [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1153.795339] env[65522]: DEBUG nova.virt.hardware [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1153.795547] env[65522]: DEBUG nova.virt.hardware [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1153.795761] env[65522]: DEBUG nova.virt.hardware [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1153.795856] env[65522]: DEBUG nova.virt.hardware [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1153.796037] env[65522]: DEBUG nova.virt.hardware [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1153.796193] env[65522]: DEBUG nova.virt.hardware [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1153.797137] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1d42ce-4324-4f6f-971a-245d5da0cb4b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.802260] env[65522]: DEBUG oslo_concurrency.lockutils [req-fe22d256-ca6b-4a9a-ba18-4850877487ee req-e5e139aa-7af5-4fab-94b6-c239c723eeaf service nova] Releasing lock "refresh_cache-59cc1b4f-b451-41d4-a071-b1ec0751bf38" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1153.805642] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5785baf7-a8df-47ad-9794-5a510c544f7b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.995644] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5267214c-2e4d-59ec-4c41-c4d9ffd7023b, 'name': SearchDatastore_Task, 'duration_secs': 0.013211} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.995807] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1153.996079] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 59cc1b4f-b451-41d4-a071-b1ec0751bf38/59cc1b4f-b451-41d4-a071-b1ec0751bf38.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1153.996424] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d034a844-efe8-4912-99e1-8d04c3a2e39f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.004571] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Waiting for the task: (returnval){ [ 1154.004571] env[65522]: value = "task-5114894" [ 1154.004571] env[65522]: _type = "Task" [ 1154.004571] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.014434] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114894, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.054215] env[65522]: WARNING openstack [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.054741] env[65522]: WARNING openstack [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.094722] env[65522]: DEBUG nova.objects.base [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Object Instance<9f342b89-bde2-4c35-ae42-cfe1e6973b74> lazy-loaded attributes: info_cache,migration_context {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1154.097393] env[65522]: INFO nova.compute.claims [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1154.103266] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c90572-23cb-46b4-8b0e-ea5b21182f95 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.126653] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f223d248-be76-4233-b0d1-3c81ea8163ee tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "99353848-2f0f-4388-9fcd-91e799342386" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.047s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.127546] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d068b1de-9299-459d-9aad-577c1f1d6503 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.134663] env[65522]: DEBUG oslo_vmware.api [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1154.134663] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523c0680-7802-9bd1-179e-800d76066205" [ 1154.134663] env[65522]: _type = "Task" [ 1154.134663] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.135614] env[65522]: DEBUG nova.network.neutron [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1154.147678] env[65522]: DEBUG oslo_vmware.api [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]523c0680-7802-9bd1-179e-800d76066205, 'name': SearchDatastore_Task, 'duration_secs': 0.009197} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.147958] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1154.156532] env[65522]: DEBUG nova.compute.manager [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1154.156532] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1154.157374] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beed88fd-e416-493d-828a-c6ca11adb9db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.166221] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1154.166567] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a409ccc-6c9c-48c4-9af2-c62dbaa15f95 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.172213] env[65522]: WARNING openstack [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.172651] env[65522]: WARNING openstack [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.185035] env[65522]: DEBUG oslo_vmware.api [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1154.185035] env[65522]: value = "task-5114895" [ 1154.185035] env[65522]: _type = "Task" [ 1154.185035] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.195176] env[65522]: DEBUG oslo_vmware.api [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114895, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.220721] env[65522]: DEBUG nova.network.neutron [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Successfully updated port: 2ff02808-9c40-4796-95c9-84dea4f51ab9 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1154.242094] env[65522]: DEBUG nova.compute.manager [req-fdae3099-a35c-477f-8c37-834267998930 req-7690cb04-ba2d-4d3e-b976-963171168ab9 service nova] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Received event network-vif-plugged-2ff02808-9c40-4796-95c9-84dea4f51ab9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1154.242094] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdae3099-a35c-477f-8c37-834267998930 req-7690cb04-ba2d-4d3e-b976-963171168ab9 service nova] Acquiring lock "af9e93f9-3e6f-4267-b7d9-1961af38c84f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1154.242094] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdae3099-a35c-477f-8c37-834267998930 req-7690cb04-ba2d-4d3e-b976-963171168ab9 service nova] Lock "af9e93f9-3e6f-4267-b7d9-1961af38c84f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1154.242094] env[65522]: DEBUG oslo_concurrency.lockutils [req-fdae3099-a35c-477f-8c37-834267998930 req-7690cb04-ba2d-4d3e-b976-963171168ab9 service nova] Lock "af9e93f9-3e6f-4267-b7d9-1961af38c84f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.243255] env[65522]: DEBUG nova.compute.manager [req-fdae3099-a35c-477f-8c37-834267998930 req-7690cb04-ba2d-4d3e-b976-963171168ab9 service nova] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] No waiting events found dispatching network-vif-plugged-2ff02808-9c40-4796-95c9-84dea4f51ab9 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1154.243255] env[65522]: WARNING nova.compute.manager [req-fdae3099-a35c-477f-8c37-834267998930 req-7690cb04-ba2d-4d3e-b976-963171168ab9 service nova] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Received unexpected event network-vif-plugged-2ff02808-9c40-4796-95c9-84dea4f51ab9 for instance with vm_state building and task_state spawning. [ 1154.255195] env[65522]: WARNING neutronclient.v2_0.client [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.255939] env[65522]: WARNING openstack [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.256363] env[65522]: WARNING openstack [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.374547] env[65522]: DEBUG nova.network.neutron [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Updating instance_info_cache with network_info: [{"id": "eb11dd2a-68ae-4ed3-a4dc-51897ede63f7", "address": "fa:16:3e:7c:70:79", "network": {"id": "2eb73cd0-541f-42e9-8774-600c5c1e9200", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-939435024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdba2dbc660145b7a98af8bd3f77a035", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb11dd2a-68", "ovs_interfaceid": "eb11dd2a-68ae-4ed3-a4dc-51897ede63f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1154.492156] env[65522]: DEBUG oslo_concurrency.lockutils [None req-af7cda66-0f08-4e15-9723-59cbb607d69c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "2eb8e092-255f-4997-8f10-a0cce0f4d503" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1154.492572] env[65522]: DEBUG oslo_concurrency.lockutils [None req-af7cda66-0f08-4e15-9723-59cbb607d69c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "2eb8e092-255f-4997-8f10-a0cce0f4d503" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1154.492893] env[65522]: DEBUG nova.compute.manager [None req-af7cda66-0f08-4e15-9723-59cbb607d69c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1154.494223] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-026b43b7-66d2-4c40-8b73-6eb5519b59e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.503265] env[65522]: DEBUG nova.compute.manager [None req-af7cda66-0f08-4e15-9723-59cbb607d69c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65522) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1154.504137] env[65522]: DEBUG nova.objects.instance [None req-af7cda66-0f08-4e15-9723-59cbb607d69c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lazy-loading 'flavor' on Instance uuid 2eb8e092-255f-4997-8f10-a0cce0f4d503 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1154.516623] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114894, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477267} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.517089] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 59cc1b4f-b451-41d4-a071-b1ec0751bf38/59cc1b4f-b451-41d4-a071-b1ec0751bf38.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1154.517351] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1154.517732] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e23eaaf-fde1-4a16-be48-ba80e805e92e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.525926] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Waiting for the task: (returnval){ [ 1154.525926] env[65522]: value = "task-5114896" [ 1154.525926] env[65522]: _type = "Task" [ 1154.525926] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.536408] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114896, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.607793] env[65522]: INFO nova.compute.resource_tracker [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating resource usage from migration 76ba5bb4-2a61-440b-a450-56d64e13a6a0 [ 1154.693596] env[65522]: DEBUG oslo_vmware.api [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114895, 'name': PowerOffVM_Task, 'duration_secs': 0.38743} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.693896] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1154.694079] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1154.694353] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-00add352-60de-45c4-ac0e-589c089feb29 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.726575] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "refresh_cache-af9e93f9-3e6f-4267-b7d9-1961af38c84f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.726817] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "refresh_cache-af9e93f9-3e6f-4267-b7d9-1961af38c84f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1154.727032] env[65522]: DEBUG nova.network.neutron [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1154.771802] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1154.772622] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1154.772622] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Deleting the datastore file [datastore2] 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1154.773013] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-781d6386-f2a3-4cd3-b4fd-e6498ebb07e1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.783071] env[65522]: DEBUG oslo_vmware.api [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1154.783071] env[65522]: value = "task-5114898" [ 1154.783071] env[65522]: _type = "Task" [ 1154.783071] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.796224] env[65522]: DEBUG oslo_vmware.api [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114898, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.833028] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744ba41a-a911-40c2-b377-ee503279916d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.841103] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a2834bb-1c59-43ad-9977-2780b201e52a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.880481] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Releasing lock "refresh_cache-b93f2d74-a24b-4f2c-aefd-16b093107ec4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1154.881250] env[65522]: DEBUG nova.compute.manager [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Instance network_info: |[{"id": "eb11dd2a-68ae-4ed3-a4dc-51897ede63f7", "address": "fa:16:3e:7c:70:79", "network": {"id": "2eb73cd0-541f-42e9-8774-600c5c1e9200", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-939435024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdba2dbc660145b7a98af8bd3f77a035", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb11dd2a-68", "ovs_interfaceid": "eb11dd2a-68ae-4ed3-a4dc-51897ede63f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1154.882323] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7c:70:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb11dd2a-68ae-4ed3-a4dc-51897ede63f7', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1154.891427] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1154.892469] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1fbc42e-6100-4d36-9b1a-15d98b1890ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.895611] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1154.896055] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3b8a6f5-1544-4e96-a1b4-62d78384f09f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.919550] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe8402e-1c20-4710-a556-4504f8d94a0c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.924773] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1154.924773] env[65522]: value = "task-5114899" [ 1154.924773] env[65522]: _type = "Task" [ 1154.924773] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.937665] env[65522]: DEBUG nova.compute.provider_tree [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1154.945455] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114899, 'name': CreateVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.037979] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114896, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088568} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.038395] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1155.039322] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-563ff797-3ef9-4b6e-8eb1-66d2e36508a7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.064621] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 59cc1b4f-b451-41d4-a071-b1ec0751bf38/59cc1b4f-b451-41d4-a071-b1ec0751bf38.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1155.065160] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3a50d11-ce17-470c-b426-8a7196e67a37 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.086583] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Waiting for the task: (returnval){ [ 1155.086583] env[65522]: value = "task-5114900" [ 1155.086583] env[65522]: _type = "Task" [ 1155.086583] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.096980] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114900, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.230479] env[65522]: WARNING openstack [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.231389] env[65522]: WARNING openstack [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.294566] env[65522]: DEBUG nova.network.neutron [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1155.303420] env[65522]: DEBUG oslo_vmware.api [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114898, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172836} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.303500] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1155.303702] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1155.303897] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1155.306651] env[65522]: INFO nova.compute.manager [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1155.306651] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1155.306651] env[65522]: DEBUG nova.compute.manager [-] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1155.306651] env[65522]: DEBUG nova.network.neutron [-] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1155.306651] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.306651] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.306651] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.314755] env[65522]: WARNING openstack [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.315184] env[65522]: WARNING openstack [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.345115] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.382015] env[65522]: WARNING neutronclient.v2_0.client [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.382893] env[65522]: WARNING openstack [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.383261] env[65522]: WARNING openstack [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.436580] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114899, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.441789] env[65522]: DEBUG nova.scheduler.client.report [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1155.498052] env[65522]: DEBUG nova.network.neutron [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Updating instance_info_cache with network_info: [{"id": "2ff02808-9c40-4796-95c9-84dea4f51ab9", "address": "fa:16:3e:26:bc:99", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ff02808-9c", "ovs_interfaceid": "2ff02808-9c40-4796-95c9-84dea4f51ab9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1155.515018] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-af7cda66-0f08-4e15-9723-59cbb607d69c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1155.515441] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3dbf070d-4d1f-40e4-aa35-05a97e73d3e0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.524815] env[65522]: DEBUG oslo_vmware.api [None req-af7cda66-0f08-4e15-9723-59cbb607d69c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1155.524815] env[65522]: value = "task-5114901" [ 1155.524815] env[65522]: _type = "Task" [ 1155.524815] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.535360] env[65522]: DEBUG oslo_vmware.api [None req-af7cda66-0f08-4e15-9723-59cbb607d69c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114901, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.601421] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114900, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.936832] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114899, 'name': CreateVM_Task, 'duration_secs': 0.803442} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.937051] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1155.937594] env[65522]: WARNING neutronclient.v2_0.client [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.938013] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.938214] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1155.938612] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1155.938806] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f72f50a0-b890-4874-b9ea-9fe79488bd73 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.944424] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1155.944424] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52172b6d-003d-a055-827d-723eb1e9491e" [ 1155.944424] env[65522]: _type = "Task" [ 1155.944424] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.948120] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.356s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1155.948308] env[65522]: INFO nova.compute.manager [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Migrating [ 1155.955631] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.506s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1155.969606] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52172b6d-003d-a055-827d-723eb1e9491e, 'name': SearchDatastore_Task, 'duration_secs': 0.010922} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.970372] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1155.970543] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1155.970819] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.970952] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1155.971159] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1155.971461] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f6184666-0a3b-40f4-b9b8-1727c861dcef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.982551] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1155.982802] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1155.983603] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb796a1d-71eb-4322-a743-aa80573aa1c2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.995228] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1155.995228] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a3aef5-0c06-c2a5-c138-5b6e2114436f" [ 1155.995228] env[65522]: _type = "Task" [ 1155.995228] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.001441] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "refresh_cache-af9e93f9-3e6f-4267-b7d9-1961af38c84f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1156.002169] env[65522]: DEBUG nova.compute.manager [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Instance network_info: |[{"id": "2ff02808-9c40-4796-95c9-84dea4f51ab9", "address": "fa:16:3e:26:bc:99", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ff02808-9c", "ovs_interfaceid": "2ff02808-9c40-4796-95c9-84dea4f51ab9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1156.005673] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:bc:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ff02808-9c40-4796-95c9-84dea4f51ab9', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1156.013840] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1156.016144] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a3aef5-0c06-c2a5-c138-5b6e2114436f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.016416] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1156.017125] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-155338c2-6c93-43bf-baba-83d07e59811e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.044773] env[65522]: DEBUG oslo_vmware.api [None req-af7cda66-0f08-4e15-9723-59cbb607d69c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114901, 'name': PowerOffVM_Task, 'duration_secs': 0.23284} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.045636] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-af7cda66-0f08-4e15-9723-59cbb607d69c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1156.045847] env[65522]: DEBUG nova.compute.manager [None req-af7cda66-0f08-4e15-9723-59cbb607d69c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1156.046143] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1156.046143] env[65522]: value = "task-5114902" [ 1156.046143] env[65522]: _type = "Task" [ 1156.046143] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.046885] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c7a1ec-215a-40e7-a7e2-9e9567067767 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.063652] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114902, 'name': CreateVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.098408] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114900, 'name': ReconfigVM_Task, 'duration_secs': 0.908098} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.099603] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 59cc1b4f-b451-41d4-a071-b1ec0751bf38/59cc1b4f-b451-41d4-a071-b1ec0751bf38.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1156.100446] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2d9b0c5-7cbb-4458-9c84-73bc25c12cbb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.110839] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Waiting for the task: (returnval){ [ 1156.110839] env[65522]: value = "task-5114903" [ 1156.110839] env[65522]: _type = "Task" [ 1156.110839] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.122469] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114903, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.288349] env[65522]: DEBUG nova.compute.manager [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Received event network-changed-2ff02808-9c40-4796-95c9-84dea4f51ab9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1156.288423] env[65522]: DEBUG nova.compute.manager [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Refreshing instance network info cache due to event network-changed-2ff02808-9c40-4796-95c9-84dea4f51ab9. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1156.289278] env[65522]: DEBUG oslo_concurrency.lockutils [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] Acquiring lock "refresh_cache-af9e93f9-3e6f-4267-b7d9-1961af38c84f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.289278] env[65522]: DEBUG oslo_concurrency.lockutils [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] Acquired lock "refresh_cache-af9e93f9-3e6f-4267-b7d9-1961af38c84f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1156.289278] env[65522]: DEBUG nova.network.neutron [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Refreshing network info cache for port 2ff02808-9c40-4796-95c9-84dea4f51ab9 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1156.313040] env[65522]: DEBUG nova.network.neutron [-] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1156.429187] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.429495] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.470951] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.471174] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1156.471411] env[65522]: DEBUG nova.network.neutron [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1156.507239] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a3aef5-0c06-c2a5-c138-5b6e2114436f, 'name': SearchDatastore_Task, 'duration_secs': 0.014069} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.508099] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f900394-df8b-4952-9c53-842c5dc68d6d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.513688] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1156.513688] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52379210-2202-6149-f562-7c2438b28690" [ 1156.513688] env[65522]: _type = "Task" [ 1156.513688] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.522511] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52379210-2202-6149-f562-7c2438b28690, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.560799] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114902, 'name': CreateVM_Task, 'duration_secs': 0.385407} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.561013] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1156.561552] env[65522]: WARNING neutronclient.v2_0.client [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1156.561939] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.562112] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1156.562427] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1156.562698] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec33813f-6816-49a3-af94-4b457f05fcdc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.567257] env[65522]: DEBUG oslo_concurrency.lockutils [None req-af7cda66-0f08-4e15-9723-59cbb607d69c tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "2eb8e092-255f-4997-8f10-a0cce0f4d503" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.075s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.570560] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1156.570560] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52501d68-9053-5606-f2b4-355d53d37fee" [ 1156.570560] env[65522]: _type = "Task" [ 1156.570560] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.579604] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52501d68-9053-5606-f2b4-355d53d37fee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.621729] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114903, 'name': Rename_Task, 'duration_secs': 0.204584} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.622989] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1156.622989] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ba82082-8119-4f50-951d-d12039a4223a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.630349] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Waiting for the task: (returnval){ [ 1156.630349] env[65522]: value = "task-5114904" [ 1156.630349] env[65522]: _type = "Task" [ 1156.630349] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.639776] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114904, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.792222] env[65522]: WARNING neutronclient.v2_0.client [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1156.793021] env[65522]: WARNING openstack [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1156.793441] env[65522]: WARNING openstack [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1156.816340] env[65522]: INFO nova.compute.manager [-] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Took 1.51 seconds to deallocate network for instance. [ 1156.896716] env[65522]: WARNING openstack [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1156.897199] env[65522]: WARNING openstack [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1156.931736] env[65522]: DEBUG nova.compute.manager [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1156.962025] env[65522]: WARNING neutronclient.v2_0.client [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1156.962800] env[65522]: WARNING openstack [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1156.963159] env[65522]: WARNING openstack [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1156.971353] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Applying migration context for instance 9f342b89-bde2-4c35-ae42-cfe1e6973b74 as it has an incoming, in-progress migration 45fe253c-dd27-4c78-85ab-b18d25cd3f71. Migration status is confirming {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1156.971667] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Applying migration context for instance cb62df50-2f86-469a-b2ee-fd3754d61c83 as it has an incoming, in-progress migration 76ba5bb4-2a61-440b-a450-56d64e13a6a0. Migration status is pre-migrating {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1156.973179] env[65522]: INFO nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating resource usage from migration 45fe253c-dd27-4c78-85ab-b18d25cd3f71 [ 1156.973501] env[65522]: INFO nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating resource usage from migration 76ba5bb4-2a61-440b-a450-56d64e13a6a0 [ 1156.976919] env[65522]: WARNING neutronclient.v2_0.client [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1156.977437] env[65522]: WARNING openstack [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1156.977776] env[65522]: WARNING openstack [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.002582] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7617189c-a902-42e7-8165-0e7c4a1de06d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1157.002582] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1157.002786] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 4aca4e87-b923-49e4-88d8-7ff51bcbe993 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1157.003408] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 2981b635-43c8-4bd6-9991-e6af0be82f3c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1157.003573] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Migration 45fe253c-dd27-4c78-85ab-b18d25cd3f71 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1157.003701] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 9f342b89-bde2-4c35-ae42-cfe1e6973b74 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1157.003815] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 3b82f9ce-ea59-430d-adb6-918cc6cc48a4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1157.003946] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 2eb8e092-255f-4997-8f10-a0cce0f4d503 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1157.004123] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 59cc1b4f-b451-41d4-a071-b1ec0751bf38 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1157.004258] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance b93f2d74-a24b-4f2c-aefd-16b093107ec4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1157.004376] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance af9e93f9-3e6f-4267-b7d9-1961af38c84f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1157.004524] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Migration 76ba5bb4-2a61-440b-a450-56d64e13a6a0 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1157.004641] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance cb62df50-2f86-469a-b2ee-fd3754d61c83 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1157.028373] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52379210-2202-6149-f562-7c2438b28690, 'name': SearchDatastore_Task, 'duration_secs': 0.012392} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.028926] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1157.029275] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] b93f2d74-a24b-4f2c-aefd-16b093107ec4/b93f2d74-a24b-4f2c-aefd-16b093107ec4.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1157.029578] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c75009d-c38e-4e35-84d9-f63004c5bd93 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.038428] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1157.038428] env[65522]: value = "task-5114905" [ 1157.038428] env[65522]: _type = "Task" [ 1157.038428] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.058768] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114905, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.082925] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52501d68-9053-5606-f2b4-355d53d37fee, 'name': SearchDatastore_Task, 'duration_secs': 0.011735} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.087023] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1157.087023] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1157.087023] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.087023] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1157.087023] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1157.087023] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-684134bf-407a-48d0-ac81-ade3ec035b84 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.095792] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1157.095998] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1157.096804] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcca4bc3-c764-4132-adbf-0a9e622dfc3f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.104409] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1157.104409] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529df2fa-f1ec-f123-a8f7-f254bf88a708" [ 1157.104409] env[65522]: _type = "Task" [ 1157.104409] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.114435] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529df2fa-f1ec-f123-a8f7-f254bf88a708, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.121402] env[65522]: DEBUG nova.network.neutron [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Updated VIF entry in instance network info cache for port 2ff02808-9c40-4796-95c9-84dea4f51ab9. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1157.121886] env[65522]: DEBUG nova.network.neutron [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Updating instance_info_cache with network_info: [{"id": "2ff02808-9c40-4796-95c9-84dea4f51ab9", "address": "fa:16:3e:26:bc:99", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ff02808-9c", "ovs_interfaceid": "2ff02808-9c40-4796-95c9-84dea4f51ab9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1157.145422] env[65522]: DEBUG oslo_vmware.api [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114904, 'name': PowerOnVM_Task, 'duration_secs': 0.470716} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.145701] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1157.145901] env[65522]: INFO nova.compute.manager [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Took 8.11 seconds to spawn the instance on the hypervisor. [ 1157.146091] env[65522]: DEBUG nova.compute.manager [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1157.146900] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a47bad-3496-42ba-9e73-5252c8ca099e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.183478] env[65522]: WARNING openstack [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.183996] env[65522]: WARNING openstack [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.254860] env[65522]: WARNING neutronclient.v2_0.client [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1157.255542] env[65522]: WARNING openstack [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.255906] env[65522]: WARNING openstack [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.323961] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1157.366864] env[65522]: DEBUG nova.network.neutron [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance_info_cache with network_info: [{"id": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "address": "fa:16:3e:5b:f2:d8", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7fc03fa-06", "ovs_interfaceid": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1157.452338] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1157.508765] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 6d856111-81bd-4e2f-b7fc-e169c3e974a6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1157.557057] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114905, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.616957] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529df2fa-f1ec-f123-a8f7-f254bf88a708, 'name': SearchDatastore_Task, 'duration_secs': 0.017319} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.617978] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd33277c-c170-4861-b66f-f01f65ee743e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.626021] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1157.626021] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52545d65-d326-cdb6-320a-f8b58dc33a1f" [ 1157.626021] env[65522]: _type = "Task" [ 1157.626021] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.629856] env[65522]: DEBUG oslo_concurrency.lockutils [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] Releasing lock "refresh_cache-af9e93f9-3e6f-4267-b7d9-1961af38c84f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1157.630124] env[65522]: DEBUG nova.compute.manager [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Received event network-vif-deleted-c2eab4eb-5a45-4c96-a145-bea8b314c82d {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1157.630385] env[65522]: INFO nova.compute.manager [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Neutron deleted interface c2eab4eb-5a45-4c96-a145-bea8b314c82d; detaching it from the instance and deleting it from the info cache [ 1157.630630] env[65522]: DEBUG nova.network.neutron [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1157.639473] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52545d65-d326-cdb6-320a-f8b58dc33a1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.665108] env[65522]: INFO nova.compute.manager [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Took 19.70 seconds to build instance. [ 1157.727907] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "2eb8e092-255f-4997-8f10-a0cce0f4d503" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1157.728303] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "2eb8e092-255f-4997-8f10-a0cce0f4d503" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1157.728618] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "2eb8e092-255f-4997-8f10-a0cce0f4d503-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1157.728893] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "2eb8e092-255f-4997-8f10-a0cce0f4d503-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1157.729196] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "2eb8e092-255f-4997-8f10-a0cce0f4d503-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1157.732030] env[65522]: INFO nova.compute.manager [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Terminating instance [ 1157.869890] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1158.012966] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance aebb7082-f88c-42f3-9d9f-93cb2a6f545b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1158.013268] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1158.013484] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3136MB phys_disk=100GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '11', 'num_vm_active': '7', 'num_task_None': '5', 'num_os_type_None': '11', 'num_proj_91597233ae9c44c094f4c32d90332fa6': '1', 'io_workload': '4', 'num_task_deleting': '1', 'num_proj_a80f979f3dc0477e9462b47f7aa87f14': '1', 'num_proj_9614f18e01594bf083e619c11760fe36': '2', 'num_vm_resized': '1', 'num_proj_0d1239b79ae94cceb89ae7a8bd57da08': '1', 'num_task_resize_prep': '1', 'num_proj_fa11b46d9fe144f391233e6eb9c819d7': '1', 'num_proj_fdba2dbc660145b7a98af8bd3f77a035': '2', 'num_proj_1546bbdca07c443d84abea1155cfde37': '1', 'num_task_powering-off': '1', 'num_vm_building': '3', 'num_task_spawning': '3', 'num_proj_367e450334d147ba9537d8c8d0be7f5d': '1', 'num_proj_f93c790ff61543bd8e134bcf9cb20bb2': '1'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1158.050847] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114905, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.667976} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.053486] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] b93f2d74-a24b-4f2c-aefd-16b093107ec4/b93f2d74-a24b-4f2c-aefd-16b093107ec4.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1158.053710] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1158.054383] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b9676f2-286a-4bbc-b85a-5284c71226a1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.061784] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1158.061784] env[65522]: value = "task-5114906" [ 1158.061784] env[65522]: _type = "Task" [ 1158.061784] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.077181] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114906, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.137134] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52545d65-d326-cdb6-320a-f8b58dc33a1f, 'name': SearchDatastore_Task, 'duration_secs': 0.045842} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.139887] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41cdf27b-1eaf-4a37-a5bf-695406e66be4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.141771] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1158.142047] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] af9e93f9-3e6f-4267-b7d9-1961af38c84f/af9e93f9-3e6f-4267-b7d9-1961af38c84f.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1158.142508] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5421a1ce-3ddc-4056-9702-28fcfc6bada1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.153029] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f6b385-a815-4f3e-a1fe-e1927eafa993 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.163561] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1158.163561] env[65522]: value = "task-5114907" [ 1158.163561] env[65522]: _type = "Task" [ 1158.163561] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.166430] env[65522]: DEBUG oslo_concurrency.lockutils [None req-03b61deb-7072-4e56-aecc-24bff985e0df tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Lock "59cc1b4f-b451-41d4-a071-b1ec0751bf38" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.212s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1158.193984] env[65522]: DEBUG nova.compute.manager [req-a13a988e-1fea-4c44-ad81-1ee057745ffc req-2196097b-97e2-4d7e-b48f-5a58dea79dab service nova] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Detach interface failed, port_id=c2eab4eb-5a45-4c96-a145-bea8b314c82d, reason: Instance 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1158.236971] env[65522]: DEBUG nova.compute.manager [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1158.237218] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1158.238211] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d0d941-26c3-4fcf-ac0f-20e89c89d14f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.248727] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1158.249071] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc41ec1f-c66b-4064-9d59-8678914ee6ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.270616] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd5ccbc-a7c9-4950-b9e5-9ad4dd61d6a3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.279066] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-403b6e86-7c60-4088-93cf-9f68a14fa8a9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.311671] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e6ebcc-8cd3-49c5-b1e2-e67cc61bc027 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.321508] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bec4b2b-559f-4537-b25f-4f59aeb350a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.327429] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1158.327718] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1158.327970] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleting the datastore file [datastore2] 2eb8e092-255f-4997-8f10-a0cce0f4d503 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1158.328759] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d5ad7233-f812-4e32-a2ad-735188e2fc33 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.340785] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1158.343595] env[65522]: DEBUG oslo_vmware.api [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1158.343595] env[65522]: value = "task-5114909" [ 1158.343595] env[65522]: _type = "Task" [ 1158.343595] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.352354] env[65522]: DEBUG oslo_vmware.api [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114909, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.578359] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114906, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07057} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.578983] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1158.579955] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0e3dbab-3b21-4ac5-ad40-5ea3f93a14dc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.616405] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] b93f2d74-a24b-4f2c-aefd-16b093107ec4/b93f2d74-a24b-4f2c-aefd-16b093107ec4.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1158.616816] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c4e49fd-e44d-4195-bb4c-04cb6b9d2841 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.644533] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1158.644533] env[65522]: value = "task-5114910" [ 1158.644533] env[65522]: _type = "Task" [ 1158.644533] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.657826] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114910, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.676786] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114907, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.845403] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1158.858189] env[65522]: DEBUG oslo_vmware.api [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114909, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.505811} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.858464] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1158.858646] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1158.858804] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1158.860085] env[65522]: INFO nova.compute.manager [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1158.860085] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1158.860085] env[65522]: DEBUG nova.compute.manager [-] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1158.860085] env[65522]: DEBUG nova.network.neutron [-] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1158.860085] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1158.860480] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.860741] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.899072] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1159.130386] env[65522]: DEBUG nova.compute.manager [req-5b3bb97f-c974-4983-bd61-29d06235ac71 req-feea6b4a-68be-4a5f-af2f-42193dbe2edd service nova] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Received event network-vif-deleted-35aba130-3bb2-45e1-9593-b70b82343229 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1159.130738] env[65522]: INFO nova.compute.manager [req-5b3bb97f-c974-4983-bd61-29d06235ac71 req-feea6b4a-68be-4a5f-af2f-42193dbe2edd service nova] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Neutron deleted interface 35aba130-3bb2-45e1-9593-b70b82343229; detaching it from the instance and deleting it from the info cache [ 1159.130738] env[65522]: DEBUG nova.network.neutron [req-5b3bb97f-c974-4983-bd61-29d06235ac71 req-feea6b4a-68be-4a5f-af2f-42193dbe2edd service nova] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1159.156742] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114910, 'name': ReconfigVM_Task, 'duration_secs': 0.413395} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.157034] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Reconfigured VM instance instance-0000006c to attach disk [datastore1] b93f2d74-a24b-4f2c-aefd-16b093107ec4/b93f2d74-a24b-4f2c-aefd-16b093107ec4.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1159.157689] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5ed1872c-e999-433f-acfa-6cf6d0ef9323 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.163801] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1159.163801] env[65522]: value = "task-5114911" [ 1159.163801] env[65522]: _type = "Task" [ 1159.163801] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.174825] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114911, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.180283] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114907, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.568549} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.180542] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] af9e93f9-3e6f-4267-b7d9-1961af38c84f/af9e93f9-3e6f-4267-b7d9-1961af38c84f.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1159.180759] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1159.181039] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-89b1324a-8f5d-4fae-a793-7fd50261b80c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.188596] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1159.188596] env[65522]: value = "task-5114912" [ 1159.188596] env[65522]: _type = "Task" [ 1159.188596] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.200992] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114912, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.354042] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1159.354348] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.399s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1159.354809] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.885s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.356563] env[65522]: INFO nova.compute.claims [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1159.359378] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1159.359556] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Cleaning up deleted instances with incomplete migration {{(pid=65522) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11971}} [ 1159.386850] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f26235b4-2ce0-4211-9e7a-6a7ae6de41b6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.408444] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance 'cb62df50-2f86-469a-b2ee-fd3754d61c83' progress to 0 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1159.608617] env[65522]: DEBUG nova.network.neutron [-] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1159.634561] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-43df8d05-9c65-4bd7-a2e1-539700368eb2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.646809] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd24c071-5c06-47b5-b05d-34382d739ce9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.674812] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114911, 'name': Rename_Task, 'duration_secs': 0.151224} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.685573] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1159.685952] env[65522]: DEBUG nova.compute.manager [req-5b3bb97f-c974-4983-bd61-29d06235ac71 req-feea6b4a-68be-4a5f-af2f-42193dbe2edd service nova] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Detach interface failed, port_id=35aba130-3bb2-45e1-9593-b70b82343229, reason: Instance 2eb8e092-255f-4997-8f10-a0cce0f4d503 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1159.686376] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a238254c-cd26-4769-9f1f-3d1de34b479a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.694748] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1159.694748] env[65522]: value = "task-5114913" [ 1159.694748] env[65522]: _type = "Task" [ 1159.694748] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.701824] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114912, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074114} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.702042] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1159.702819] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72bbad7-c720-4e11-91cd-3d5254762d39 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.708421] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114913, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.729265] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] af9e93f9-3e6f-4267-b7d9-1961af38c84f/af9e93f9-3e6f-4267-b7d9-1961af38c84f.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1159.729606] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94256e6b-a06b-4e24-8082-e74ae0406456 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.749593] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1159.749593] env[65522]: value = "task-5114914" [ 1159.749593] env[65522]: _type = "Task" [ 1159.749593] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.760902] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114914, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.865556] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1159.914826] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1159.915220] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c52edb01-f55b-4b35-ba09-091187da0331 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.923545] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1159.923545] env[65522]: value = "task-5114915" [ 1159.923545] env[65522]: _type = "Task" [ 1159.923545] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.933309] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Acquiring lock "59cc1b4f-b451-41d4-a071-b1ec0751bf38" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1159.933691] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Lock "59cc1b4f-b451-41d4-a071-b1ec0751bf38" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.933990] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Acquiring lock "59cc1b4f-b451-41d4-a071-b1ec0751bf38-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1159.934230] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Lock "59cc1b4f-b451-41d4-a071-b1ec0751bf38-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.934440] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Lock "59cc1b4f-b451-41d4-a071-b1ec0751bf38-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1159.936240] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114915, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.936904] env[65522]: INFO nova.compute.manager [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Terminating instance [ 1160.111073] env[65522]: INFO nova.compute.manager [-] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Took 1.25 seconds to deallocate network for instance. [ 1160.206735] env[65522]: DEBUG oslo_vmware.api [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114913, 'name': PowerOnVM_Task, 'duration_secs': 0.503191} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.207223] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1160.207393] env[65522]: INFO nova.compute.manager [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Took 8.80 seconds to spawn the instance on the hypervisor. [ 1160.207607] env[65522]: DEBUG nova.compute.manager [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1160.208645] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9651c446-aca0-4dce-aa65-4019c106c1a6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.260271] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114914, 'name': ReconfigVM_Task, 'duration_secs': 0.325322} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.260553] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Reconfigured VM instance instance-0000006d to attach disk [datastore1] af9e93f9-3e6f-4267-b7d9-1961af38c84f/af9e93f9-3e6f-4267-b7d9-1961af38c84f.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1160.261214] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f93e408-9d33-492a-a0a7-8844261b1408 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.268765] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1160.268765] env[65522]: value = "task-5114916" [ 1160.268765] env[65522]: _type = "Task" [ 1160.268765] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.278940] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114916, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.435499] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114915, 'name': PowerOffVM_Task, 'duration_secs': 0.40005} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.435778] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1160.435961] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance 'cb62df50-2f86-469a-b2ee-fd3754d61c83' progress to 17 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1160.441487] env[65522]: DEBUG nova.compute.manager [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1160.441686] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1160.442495] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568bc3f6-f740-42ec-8f6c-9f1a6e2d46d4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.449891] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1160.450303] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7eb834df-3fc0-4dc1-8db2-85f84170d06c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.457145] env[65522]: DEBUG oslo_vmware.api [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Waiting for the task: (returnval){ [ 1160.457145] env[65522]: value = "task-5114917" [ 1160.457145] env[65522]: _type = "Task" [ 1160.457145] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.466352] env[65522]: DEBUG oslo_vmware.api [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114917, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.618099] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1160.668485] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc32dec3-0d76-47f0-a5db-37076ce025cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.676741] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc01a91-699a-404a-bc41-a5a2e1c26569 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.709974] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535b5bd9-8392-4d97-a93f-e4f187e8cff4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.718367] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8557a69f-1ca5-448a-b4ce-49addc2577f0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.729632] env[65522]: INFO nova.compute.manager [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Took 20.34 seconds to build instance. [ 1160.738833] env[65522]: DEBUG nova.compute.provider_tree [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1160.779340] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114916, 'name': Rename_Task, 'duration_secs': 0.289478} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.779622] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1160.780231] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ec115f0-f514-4093-9375-05bd33df6c49 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.786898] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1160.786898] env[65522]: value = "task-5114918" [ 1160.786898] env[65522]: _type = "Task" [ 1160.786898] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.795117] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114918, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.942272] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1160.942528] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1160.943141] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1160.943141] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1160.943141] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1160.943314] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1160.943424] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1160.943581] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1160.943740] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1160.943957] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1160.944163] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1160.949688] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7be5837b-4271-43ae-a7a7-86a5cbbb6978 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.971224] env[65522]: DEBUG oslo_vmware.api [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114917, 'name': PowerOffVM_Task, 'duration_secs': 0.199852} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.972568] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1160.972743] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1160.973107] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1160.973107] env[65522]: value = "task-5114919" [ 1160.973107] env[65522]: _type = "Task" [ 1160.973107] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.973318] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e387c19-60b4-4307-8bfa-9c3cf9f4a94f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.984535] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114919, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.050583] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1161.050827] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1161.051013] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Deleting the datastore file [datastore1] 59cc1b4f-b451-41d4-a071-b1ec0751bf38 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1161.051312] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f642d8bf-19b4-481b-a8a9-c31ad865f575 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.059899] env[65522]: DEBUG oslo_vmware.api [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Waiting for the task: (returnval){ [ 1161.059899] env[65522]: value = "task-5114921" [ 1161.059899] env[65522]: _type = "Task" [ 1161.059899] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.072802] env[65522]: DEBUG oslo_vmware.api [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114921, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.162480] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.162760] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.241880] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8a6d1b23-c263-4cfa-a5f9-b0b20ea93a8f tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "b93f2d74-a24b-4f2c-aefd-16b093107ec4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.860s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1161.243068] env[65522]: DEBUG nova.scheduler.client.report [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1161.298445] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114918, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.402332] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "b93f2d74-a24b-4f2c-aefd-16b093107ec4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1161.402524] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "b93f2d74-a24b-4f2c-aefd-16b093107ec4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1161.402728] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "b93f2d74-a24b-4f2c-aefd-16b093107ec4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1161.402965] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "b93f2d74-a24b-4f2c-aefd-16b093107ec4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1161.403149] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "b93f2d74-a24b-4f2c-aefd-16b093107ec4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1161.405399] env[65522]: INFO nova.compute.manager [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Terminating instance [ 1161.486050] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114919, 'name': ReconfigVM_Task, 'duration_secs': 0.229205} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.486464] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance 'cb62df50-2f86-469a-b2ee-fd3754d61c83' progress to 33 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1161.571883] env[65522]: DEBUG oslo_vmware.api [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Task: {'id': task-5114921, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162079} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.572142] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1161.572322] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1161.572489] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1161.572653] env[65522]: INFO nova.compute.manager [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1161.572920] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1161.573144] env[65522]: DEBUG nova.compute.manager [-] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1161.573242] env[65522]: DEBUG nova.network.neutron [-] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1161.573477] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1161.574067] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1161.574323] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1161.611051] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1161.669622] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.670049] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.670049] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.670150] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.670262] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.670408] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.670539] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1161.670701] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.747713] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1161.748352] env[65522]: DEBUG nova.compute.manager [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1161.750928] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 7.603s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1161.799068] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114918, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.886034] env[65522]: DEBUG nova.compute.manager [req-708f9e91-a72b-4b8a-86c6-c6d6e4243e94 req-53ebcbeb-3f53-4ecc-9fc5-c3e61daf3c9a service nova] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Received event network-vif-deleted-ca385019-1696-4857-b5cc-bc42db6d70f4 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1161.886493] env[65522]: INFO nova.compute.manager [req-708f9e91-a72b-4b8a-86c6-c6d6e4243e94 req-53ebcbeb-3f53-4ecc-9fc5-c3e61daf3c9a service nova] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Neutron deleted interface ca385019-1696-4857-b5cc-bc42db6d70f4; detaching it from the instance and deleting it from the info cache [ 1161.886493] env[65522]: DEBUG nova.network.neutron [req-708f9e91-a72b-4b8a-86c6-c6d6e4243e94 req-53ebcbeb-3f53-4ecc-9fc5-c3e61daf3c9a service nova] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1161.909364] env[65522]: DEBUG nova.compute.manager [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1161.909588] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1161.911148] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233df37d-a52a-4932-a446-2d0055f8c14c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.920690] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1161.920970] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3782f756-ae94-4886-93e3-a6e67d4b501a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.928467] env[65522]: DEBUG oslo_vmware.api [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1161.928467] env[65522]: value = "task-5114922" [ 1161.928467] env[65522]: _type = "Task" [ 1161.928467] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.936800] env[65522]: DEBUG oslo_vmware.api [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114922, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.993447] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1161.993711] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1161.993891] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1161.994125] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1161.994274] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1161.994419] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1161.994623] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1161.994783] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1161.994950] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1161.995141] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1161.995359] env[65522]: DEBUG nova.virt.hardware [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1162.001058] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Reconfiguring VM instance instance-00000064 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1162.001421] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da22f7a0-92fd-4e9d-88fd-92cefef60a94 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.022318] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1162.022318] env[65522]: value = "task-5114923" [ 1162.022318] env[65522]: _type = "Task" [ 1162.022318] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.031910] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114923, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.173240] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1162.254244] env[65522]: DEBUG nova.compute.utils [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1162.258533] env[65522]: DEBUG nova.compute.manager [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Not allocating networking since 'none' was specified. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 1162.303103] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114918, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.341773] env[65522]: DEBUG nova.network.neutron [-] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1162.390056] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8beda987-cb26-4395-820d-1ac87a5358db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.400894] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ae1cfd-f8a1-41e3-8a30-0aa0654aaa8b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.439493] env[65522]: DEBUG nova.compute.manager [req-708f9e91-a72b-4b8a-86c6-c6d6e4243e94 req-53ebcbeb-3f53-4ecc-9fc5-c3e61daf3c9a service nova] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Detach interface failed, port_id=ca385019-1696-4857-b5cc-bc42db6d70f4, reason: Instance 59cc1b4f-b451-41d4-a071-b1ec0751bf38 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1162.449635] env[65522]: DEBUG oslo_vmware.api [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114922, 'name': PowerOffVM_Task, 'duration_secs': 0.184055} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.452634] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1162.452886] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1162.453438] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-111542ba-e1bb-4a0f-9c46-384d21e89f92 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.507188] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a10084a2-3e91-41ea-8340-4983fc94f893 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.516469] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d91aaab9-6b97-48a2-8b6c-b7cfbc10dd5b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.553442] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52c12d5-2bd3-48cd-be30-0ef9ead1ba31 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.555849] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1162.556059] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1162.556238] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Deleting the datastore file [datastore1] b93f2d74-a24b-4f2c-aefd-16b093107ec4 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1162.556647] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b50dca3c-89dc-4bb1-8760-ce4ad3102a6a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.567374] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bffe9316-5240-4160-8000-6b2308005896 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.571479] env[65522]: DEBUG oslo_vmware.api [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1162.571479] env[65522]: value = "task-5114925" [ 1162.571479] env[65522]: _type = "Task" [ 1162.571479] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.571814] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114923, 'name': ReconfigVM_Task, 'duration_secs': 0.195248} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.572247] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Reconfigured VM instance instance-00000064 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1162.576948] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8837316e-0de9-4738-8075-da6237da8e92 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.586978] env[65522]: DEBUG nova.compute.provider_tree [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1162.594212] env[65522]: DEBUG oslo_vmware.api [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114925, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.613717] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] cb62df50-2f86-469a-b2ee-fd3754d61c83/cb62df50-2f86-469a-b2ee-fd3754d61c83.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1162.614688] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc86ab44-e604-4099-91fb-5c7ffe9cfbc0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.633939] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1162.633939] env[65522]: value = "task-5114926" [ 1162.633939] env[65522]: _type = "Task" [ 1162.633939] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.643429] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114926, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.760013] env[65522]: DEBUG nova.compute.manager [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1162.799131] env[65522]: DEBUG oslo_vmware.api [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114918, 'name': PowerOnVM_Task, 'duration_secs': 1.574531} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.799423] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1162.799627] env[65522]: INFO nova.compute.manager [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Took 9.03 seconds to spawn the instance on the hypervisor. [ 1162.799810] env[65522]: DEBUG nova.compute.manager [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1162.800627] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09925e01-0c71-44ff-aba3-476c07dc6735 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.844692] env[65522]: INFO nova.compute.manager [-] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Took 1.27 seconds to deallocate network for instance. [ 1163.083039] env[65522]: DEBUG oslo_vmware.api [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114925, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155458} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.083039] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1163.083039] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1163.083039] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1163.083298] env[65522]: INFO nova.compute.manager [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1163.083351] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1163.083556] env[65522]: DEBUG nova.compute.manager [-] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1163.083656] env[65522]: DEBUG nova.network.neutron [-] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1163.083930] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1163.084509] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1163.084765] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1163.092023] env[65522]: DEBUG nova.scheduler.client.report [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1163.145179] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114926, 'name': ReconfigVM_Task, 'duration_secs': 0.449769} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.145361] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Reconfigured VM instance instance-00000064 to attach disk [datastore2] cb62df50-2f86-469a-b2ee-fd3754d61c83/cb62df50-2f86-469a-b2ee-fd3754d61c83.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1163.145634] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance 'cb62df50-2f86-469a-b2ee-fd3754d61c83' progress to 50 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1163.153205] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1163.319722] env[65522]: INFO nova.compute.manager [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Took 18.26 seconds to build instance. [ 1163.354818] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1163.652316] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70a1381-eb33-45b9-89ac-86b88806e047 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.674021] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd9d9e9-2318-4902-a959-9b98c4c45aeb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.692398] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance 'cb62df50-2f86-469a-b2ee-fd3754d61c83' progress to 67 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1163.697987] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be74898-17f3-4481-b402-2eebc92817e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.704334] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-97cd2a4d-5d25-48fa-99c2-9dc833c34ced tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Suspending the VM {{(pid=65522) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1163.704580] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-5aa00c08-7146-4ecd-a25c-c26f35526ce7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.712015] env[65522]: DEBUG oslo_vmware.api [None req-97cd2a4d-5d25-48fa-99c2-9dc833c34ced tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1163.712015] env[65522]: value = "task-5114927" [ 1163.712015] env[65522]: _type = "Task" [ 1163.712015] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.721274] env[65522]: DEBUG oslo_vmware.api [None req-97cd2a4d-5d25-48fa-99c2-9dc833c34ced tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114927, 'name': SuspendVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.769179] env[65522]: DEBUG nova.compute.manager [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1163.798024] env[65522]: DEBUG nova.virt.hardware [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1163.798024] env[65522]: DEBUG nova.virt.hardware [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1163.798024] env[65522]: DEBUG nova.virt.hardware [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1163.798411] env[65522]: DEBUG nova.virt.hardware [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1163.798734] env[65522]: DEBUG nova.virt.hardware [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1163.800058] env[65522]: DEBUG nova.virt.hardware [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1163.800058] env[65522]: DEBUG nova.virt.hardware [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1163.800058] env[65522]: DEBUG nova.virt.hardware [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1163.800058] env[65522]: DEBUG nova.virt.hardware [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1163.800058] env[65522]: DEBUG nova.virt.hardware [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1163.800058] env[65522]: DEBUG nova.virt.hardware [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1163.800947] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7fffc80-87b8-472d-b33e-108d70202bba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.809637] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35893c7e-8f8d-435b-9394-0b4685b85660 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.824134] env[65522]: DEBUG oslo_concurrency.lockutils [None req-82e4436a-65ae-485b-8046-cff6b5a36b1f tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "af9e93f9-3e6f-4267-b7d9-1961af38c84f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.770s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1163.824697] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1163.830277] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Creating folder: Project (ad195a7fa9cc4c2e8cad7ddcc5f0f62a). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1163.830626] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3783bdab-0a1b-46b9-bf38-5f38dc44f4c8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.843081] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Created folder: Project (ad195a7fa9cc4c2e8cad7ddcc5f0f62a) in parent group-v994660. [ 1163.843325] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Creating folder: Instances. Parent ref: group-v994951. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1163.843557] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0fd7f5fc-73fd-4f40-bec0-892faee0d001 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.854198] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Created folder: Instances in parent group-v994951. [ 1163.854462] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1163.854671] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1163.854888] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f9355794-6acc-4cd4-ad99-deb986d6e0b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.873382] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1163.873382] env[65522]: value = "task-5114930" [ 1163.873382] env[65522]: _type = "Task" [ 1163.873382] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.881738] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114930, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.883194] env[65522]: DEBUG nova.network.neutron [-] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1163.920412] env[65522]: DEBUG nova.compute.manager [req-01a66664-8878-4fbe-a206-2030a99a6309 req-c6681191-b608-4b69-aa58-5705cdeb403b service nova] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Received event network-vif-deleted-eb11dd2a-68ae-4ed3-a4dc-51897ede63f7 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1164.104310] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.353s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.104557] env[65522]: DEBUG nova.compute.manager [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=65522) _confirm_resize /opt/stack/nova/nova/compute/manager.py:5433}} [ 1164.107517] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.784s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1164.108804] env[65522]: DEBUG nova.objects.instance [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lazy-loading 'resources' on Instance uuid 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.200793] env[65522]: WARNING neutronclient.v2_0.client [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1164.227806] env[65522]: DEBUG oslo_vmware.api [None req-97cd2a4d-5d25-48fa-99c2-9dc833c34ced tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114927, 'name': SuspendVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.250433] env[65522]: DEBUG nova.network.neutron [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Port a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac binding to destination host cpu-1 is already ACTIVE {{(pid=65522) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 1164.384098] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114930, 'name': CreateVM_Task, 'duration_secs': 0.365139} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.384291] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1164.384718] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.384880] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1164.385279] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1164.385737] env[65522]: INFO nova.compute.manager [-] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Took 1.30 seconds to deallocate network for instance. [ 1164.385959] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6a4b8a5-d34f-49e8-8d2e-da712c840f70 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.393223] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1164.393223] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52889170-8778-741e-2a05-1f45337801b8" [ 1164.393223] env[65522]: _type = "Task" [ 1164.393223] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.403350] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52889170-8778-741e-2a05-1f45337801b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.668868] env[65522]: INFO nova.scheduler.client.report [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleted allocation for migration 45fe253c-dd27-4c78-85ab-b18d25cd3f71 [ 1164.723178] env[65522]: DEBUG oslo_vmware.api [None req-97cd2a4d-5d25-48fa-99c2-9dc833c34ced tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114927, 'name': SuspendVM_Task, 'duration_secs': 0.67097} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.723419] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-97cd2a4d-5d25-48fa-99c2-9dc833c34ced tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Suspended the VM {{(pid=65522) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1164.723563] env[65522]: DEBUG nova.compute.manager [None req-97cd2a4d-5d25-48fa-99c2-9dc833c34ced tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1164.724764] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf21e6b-837a-4c7f-bfb6-3d989805b776 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.805378] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4ceb3f-f4bb-4337-ac4f-7389921cc5ca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.813064] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e911d4-314f-4646-84d7-347a159776b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.844773] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87dc89f2-68bd-4c02-9980-051a09f28ab5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.852912] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f525fb-a5bd-4d2d-b402-907d63804ebc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.866731] env[65522]: DEBUG nova.compute.provider_tree [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1164.895884] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1164.905678] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52889170-8778-741e-2a05-1f45337801b8, 'name': SearchDatastore_Task, 'duration_secs': 0.011257} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.905968] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1164.906217] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1164.906449] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.906592] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1164.906767] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1164.907329] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb252356-a923-4bb9-9c46-6f476c1d99a9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.915846] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1164.916029] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1164.916734] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-849f81eb-beca-4d7e-9a70-90c675e62f0b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.921700] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1164.921700] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d3c171-6da2-c7ba-f4f1-6c928f59c605" [ 1164.921700] env[65522]: _type = "Task" [ 1164.921700] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.929339] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d3c171-6da2-c7ba-f4f1-6c928f59c605, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.005373] env[65522]: DEBUG nova.objects.instance [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'flavor' on Instance uuid 9f342b89-bde2-4c35-ae42-cfe1e6973b74 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.176821] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5048f45d-d42a-42ae-8163-2a891e41c49a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 14.291s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1165.275346] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "cb62df50-2f86-469a-b2ee-fd3754d61c83-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1165.276061] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "cb62df50-2f86-469a-b2ee-fd3754d61c83-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1165.276061] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "cb62df50-2f86-469a-b2ee-fd3754d61c83-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1165.369795] env[65522]: DEBUG nova.scheduler.client.report [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1165.432595] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d3c171-6da2-c7ba-f4f1-6c928f59c605, 'name': SearchDatastore_Task, 'duration_secs': 0.00843} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.433458] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-228569b4-af78-45c2-abf3-9e2a6f28a0db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.438928] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1165.438928] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52247bc5-40af-2c19-e29c-e23fbd3c5936" [ 1165.438928] env[65522]: _type = "Task" [ 1165.438928] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.449473] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52247bc5-40af-2c19-e29c-e23fbd3c5936, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.510447] env[65522]: DEBUG oslo_concurrency.lockutils [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.510447] env[65522]: DEBUG oslo_concurrency.lockutils [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1165.510606] env[65522]: DEBUG nova.network.neutron [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1165.510767] env[65522]: DEBUG nova.objects.instance [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'info_cache' on Instance uuid 9f342b89-bde2-4c35-ae42-cfe1e6973b74 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.874568] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.767s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1165.877070] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.425s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1165.879200] env[65522]: INFO nova.compute.claims [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1165.897071] env[65522]: INFO nova.scheduler.client.report [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Deleted allocations for instance 5ce4a286-efd1-4bbc-a23b-931c6701cfe4 [ 1165.949876] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52247bc5-40af-2c19-e29c-e23fbd3c5936, 'name': SearchDatastore_Task, 'duration_secs': 0.010937} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.950618] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1165.950902] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 6d856111-81bd-4e2f-b7fc-e169c3e974a6/6d856111-81bd-4e2f-b7fc-e169c3e974a6.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1165.951192] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-81737c86-b3ac-4a78-83ba-0dafcc336747 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.958709] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1165.958709] env[65522]: value = "task-5114931" [ 1165.958709] env[65522]: _type = "Task" [ 1165.958709] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.967334] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114931, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.017682] env[65522]: DEBUG nova.objects.base [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Object Instance<9f342b89-bde2-4c35-ae42-cfe1e6973b74> lazy-loaded attributes: flavor,info_cache {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1166.249024] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "af9e93f9-3e6f-4267-b7d9-1961af38c84f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1166.249399] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "af9e93f9-3e6f-4267-b7d9-1961af38c84f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1166.249687] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "af9e93f9-3e6f-4267-b7d9-1961af38c84f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1166.249919] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "af9e93f9-3e6f-4267-b7d9-1961af38c84f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1166.250121] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "af9e93f9-3e6f-4267-b7d9-1961af38c84f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.252368] env[65522]: INFO nova.compute.manager [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Terminating instance [ 1166.282044] env[65522]: WARNING neutronclient.v2_0.client [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.316820] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.317084] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1166.317276] env[65522]: DEBUG nova.network.neutron [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1166.406544] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c71f99d8-cf5f-4c4b-9ce9-c23e4e5c901a tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "5ce4a286-efd1-4bbc-a23b-931c6701cfe4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.758s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.469292] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114931, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.521825] env[65522]: WARNING neutronclient.v2_0.client [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.522670] env[65522]: WARNING openstack [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.522720] env[65522]: WARNING openstack [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.674124] env[65522]: WARNING openstack [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.674448] env[65522]: WARNING openstack [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.734943] env[65522]: WARNING neutronclient.v2_0.client [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.735639] env[65522]: WARNING openstack [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.736873] env[65522]: WARNING openstack [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.756997] env[65522]: DEBUG nova.compute.manager [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1166.756997] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1166.757867] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb41af9-5928-47c1-bb86-157caa47f7e2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.767275] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1166.767577] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-36a0ed24-58f1-447f-b8bf-8f0a614ada4e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.820092] env[65522]: WARNING neutronclient.v2_0.client [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.820949] env[65522]: WARNING openstack [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.821370] env[65522]: WARNING openstack [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.839207] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1166.839364] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1166.839551] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleting the datastore file [datastore1] af9e93f9-3e6f-4267-b7d9-1961af38c84f {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1166.842291] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6aa06c7f-dfae-4db6-83ee-488f2c8faa43 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.850696] env[65522]: DEBUG oslo_vmware.api [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1166.850696] env[65522]: value = "task-5114933" [ 1166.850696] env[65522]: _type = "Task" [ 1166.850696] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.856951] env[65522]: DEBUG nova.network.neutron [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance_info_cache with network_info: [{"id": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "address": "fa:16:3e:c3:12:7d", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped6e9224-69", "ovs_interfaceid": "ed6e9224-69f8-46fd-a71f-27d8bc6eaec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1166.864567] env[65522]: DEBUG oslo_vmware.api [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114933, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.970246] env[65522]: WARNING openstack [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.970927] env[65522]: WARNING openstack [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.984039] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114931, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.552035} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.984305] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 6d856111-81bd-4e2f-b7fc-e169c3e974a6/6d856111-81bd-4e2f-b7fc-e169c3e974a6.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1166.984619] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1166.984775] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f7966bd3-150f-4695-ac47-0c1e9ef7d4b6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.994699] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1166.994699] env[65522]: value = "task-5114934" [ 1166.994699] env[65522]: _type = "Task" [ 1166.994699] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.007833] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114934, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.080456] env[65522]: WARNING neutronclient.v2_0.client [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1167.081471] env[65522]: WARNING openstack [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1167.082234] env[65522]: WARNING openstack [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1167.104279] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad78e6b-b85b-4e10-8017-27a0becee7a3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.112526] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864dbb8e-c6af-4041-8f87-911e62452630 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.148623] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98f8922-d4ac-45e8-b623-11e171529f3f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.157280] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bacca377-bc8d-4b47-90f2-a36e487263dd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.171356] env[65522]: DEBUG nova.compute.provider_tree [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.188152] env[65522]: DEBUG nova.network.neutron [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance_info_cache with network_info: [{"id": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "address": "fa:16:3e:5b:f2:d8", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7fc03fa-06", "ovs_interfaceid": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1167.361534] env[65522]: DEBUG oslo_concurrency.lockutils [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "refresh_cache-9f342b89-bde2-4c35-ae42-cfe1e6973b74" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1167.363335] env[65522]: DEBUG oslo_vmware.api [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114933, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150316} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.363827] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1167.364060] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1167.364245] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1167.364412] env[65522]: INFO nova.compute.manager [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1167.364654] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1167.364846] env[65522]: DEBUG nova.compute.manager [-] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1167.364944] env[65522]: DEBUG nova.network.neutron [-] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1167.365268] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1167.365792] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1167.366060] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1167.403393] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1167.504910] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114934, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084518} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.505293] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1167.506165] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d5fc64-a371-4637-9b30-66e8b20db131 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.529019] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 6d856111-81bd-4e2f-b7fc-e169c3e974a6/6d856111-81bd-4e2f-b7fc-e169c3e974a6.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1167.529019] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01860402-7a0f-4a9f-a576-38494b85045c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.549339] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1167.549339] env[65522]: value = "task-5114935" [ 1167.549339] env[65522]: _type = "Task" [ 1167.549339] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.558587] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114935, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.675318] env[65522]: DEBUG nova.scheduler.client.report [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1167.690801] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1167.760793] env[65522]: DEBUG nova.compute.manager [req-7ba8fb2f-3dcb-43c5-a751-649eaaad59c9 req-5fe206ba-ba21-4777-8d0a-19f50356c621 service nova] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Received event network-vif-deleted-2ff02808-9c40-4796-95c9-84dea4f51ab9 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1167.761065] env[65522]: INFO nova.compute.manager [req-7ba8fb2f-3dcb-43c5-a751-649eaaad59c9 req-5fe206ba-ba21-4777-8d0a-19f50356c621 service nova] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Neutron deleted interface 2ff02808-9c40-4796-95c9-84dea4f51ab9; detaching it from the instance and deleting it from the info cache [ 1167.761190] env[65522]: DEBUG nova.network.neutron [req-7ba8fb2f-3dcb-43c5-a751-649eaaad59c9 req-5fe206ba-ba21-4777-8d0a-19f50356c621 service nova] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1168.059466] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114935, 'name': ReconfigVM_Task, 'duration_secs': 0.308446} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.059924] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 6d856111-81bd-4e2f-b7fc-e169c3e974a6/6d856111-81bd-4e2f-b7fc-e169c3e974a6.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1168.060416] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d42ca446-bf86-45a4-a2ab-1035cdeaa0cb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.068232] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1168.068232] env[65522]: value = "task-5114936" [ 1168.068232] env[65522]: _type = "Task" [ 1168.068232] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.078342] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114936, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.180499] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.303s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.181132] env[65522]: DEBUG nova.compute.manager [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1168.183782] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.566s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1168.184093] env[65522]: DEBUG nova.objects.instance [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lazy-loading 'resources' on Instance uuid 2eb8e092-255f-4997-8f10-a0cce0f4d503 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.213700] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19361b7a-4471-48ad-bf47-5a6d66680004 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.233779] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-070e4afc-1658-46db-b9e3-4f102fad1fa5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.238427] env[65522]: DEBUG nova.network.neutron [-] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1168.242994] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance 'cb62df50-2f86-469a-b2ee-fd3754d61c83' progress to 83 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1168.264387] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-99b595b8-d0fc-408a-bd3a-a96eb6fada81 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.274314] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abef2c50-81ba-417c-800b-6b69231865d1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.314284] env[65522]: DEBUG nova.compute.manager [req-7ba8fb2f-3dcb-43c5-a751-649eaaad59c9 req-5fe206ba-ba21-4777-8d0a-19f50356c621 service nova] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Detach interface failed, port_id=2ff02808-9c40-4796-95c9-84dea4f51ab9, reason: Instance af9e93f9-3e6f-4267-b7d9-1961af38c84f could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1168.367480] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1168.367843] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-76cbfa79-8438-4568-a043-c0b415f4a6e0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.375661] env[65522]: DEBUG oslo_vmware.api [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1168.375661] env[65522]: value = "task-5114937" [ 1168.375661] env[65522]: _type = "Task" [ 1168.375661] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.384309] env[65522]: DEBUG oslo_vmware.api [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114937, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.578821] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114936, 'name': Rename_Task, 'duration_secs': 0.179087} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.579128] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1168.579440] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a766e27b-0856-4043-b4d6-d78d5964284b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.586892] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1168.586892] env[65522]: value = "task-5114938" [ 1168.586892] env[65522]: _type = "Task" [ 1168.586892] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.595081] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114938, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.667923] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "3987d71f-47ab-4dd4-8426-95d78503effa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1168.668186] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1168.687047] env[65522]: DEBUG nova.compute.utils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1168.691246] env[65522]: DEBUG nova.compute.manager [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1168.691504] env[65522]: DEBUG nova.network.neutron [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1168.691863] env[65522]: WARNING neutronclient.v2_0.client [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1168.692297] env[65522]: WARNING neutronclient.v2_0.client [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1168.692975] env[65522]: WARNING openstack [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1168.693405] env[65522]: WARNING openstack [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1168.748987] env[65522]: INFO nova.compute.manager [-] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Took 1.38 seconds to deallocate network for instance. [ 1168.750727] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1168.754349] env[65522]: DEBUG nova.policy [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f010fe7784043aaa3d4d44b921c438f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91597233ae9c44c094f4c32d90332fa6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1168.757698] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3889fc65-fb80-4de0-ab19-bb33e69c9892 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.772122] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1168.772122] env[65522]: value = "task-5114939" [ 1168.772122] env[65522]: _type = "Task" [ 1168.772122] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.781224] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114939, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.892181] env[65522]: DEBUG oslo_vmware.api [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114937, 'name': PowerOnVM_Task, 'duration_secs': 0.415999} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.892967] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1168.892967] env[65522]: DEBUG nova.compute.manager [None req-47a7f24e-2a21-4c41-bbad-dc8bc12943fa tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1168.893484] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e8654f2-8b5e-4592-a73a-0a274e857fa8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.898260] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0288af3-30e4-4cf5-a061-edb4a78c0987 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.907865] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a689e38-25bf-4f37-b5b8-3fe763d8550e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.944132] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6cd8840-db56-4174-850f-6a14d4669dce {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.953538] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e1869b-f93d-4237-ac14-e738939c5de9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.968981] env[65522]: DEBUG nova.compute.provider_tree [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1169.055497] env[65522]: DEBUG nova.network.neutron [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Successfully created port: 0b80621b-795f-4503-9161-c9799b6339b6 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1169.098703] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114938, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.174024] env[65522]: DEBUG nova.compute.manager [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1169.191648] env[65522]: DEBUG nova.compute.manager [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1169.269561] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1169.283163] env[65522]: DEBUG oslo_vmware.api [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114939, 'name': PowerOnVM_Task, 'duration_secs': 0.425142} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.283459] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1169.283641] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55591e13-58d8-49c0-a527-cef7ea13c254 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance 'cb62df50-2f86-469a-b2ee-fd3754d61c83' progress to 100 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1169.472310] env[65522]: DEBUG nova.scheduler.client.report [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1169.601395] env[65522]: DEBUG oslo_vmware.api [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114938, 'name': PowerOnVM_Task, 'duration_secs': 0.580883} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.601681] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1169.601886] env[65522]: INFO nova.compute.manager [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Took 5.83 seconds to spawn the instance on the hypervisor. [ 1169.602084] env[65522]: DEBUG nova.compute.manager [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1169.603140] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7017f57d-c99d-4f66-b501-e78f001f750b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.700465] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1169.798195] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1169.798430] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1169.798646] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1169.798821] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1169.798984] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1169.801362] env[65522]: INFO nova.compute.manager [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Terminating instance [ 1169.978344] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.794s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1169.980691] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 7.808s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1169.980861] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1169.981038] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1169.981472] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.627s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1169.981541] env[65522]: DEBUG nova.objects.instance [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Lazy-loading 'resources' on Instance uuid 59cc1b4f-b451-41d4-a071-b1ec0751bf38 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1169.983244] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea0dc8f6-5c0d-4c92-833a-f883204ff297 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.994101] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f153f303-e3ce-45f4-86cc-be0d7c9f2758 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.999297] env[65522]: INFO nova.scheduler.client.report [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleted allocations for instance 2eb8e092-255f-4997-8f10-a0cce0f4d503 [ 1170.016107] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4028c710-b979-4fc4-b3f6-300a1f120fe0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.025070] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78529a6-daf6-4f69-8b8e-e630cf6c814e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.062470] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178537MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1170.062470] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1170.121990] env[65522]: INFO nova.compute.manager [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Took 20.67 seconds to build instance. [ 1170.203730] env[65522]: DEBUG nova.compute.manager [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1170.232809] env[65522]: DEBUG nova.virt.hardware [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1170.233141] env[65522]: DEBUG nova.virt.hardware [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1170.233302] env[65522]: DEBUG nova.virt.hardware [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1170.233483] env[65522]: DEBUG nova.virt.hardware [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1170.233625] env[65522]: DEBUG nova.virt.hardware [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1170.233769] env[65522]: DEBUG nova.virt.hardware [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1170.234012] env[65522]: DEBUG nova.virt.hardware [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1170.234197] env[65522]: DEBUG nova.virt.hardware [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1170.234367] env[65522]: DEBUG nova.virt.hardware [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1170.234526] env[65522]: DEBUG nova.virt.hardware [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1170.234697] env[65522]: DEBUG nova.virt.hardware [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1170.235638] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17795794-b7cc-4fa0-9c9b-426316173b37 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.245465] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ffab7e1-50bf-4e00-b1b2-fd97bc5b8691 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.304800] env[65522]: DEBUG nova.compute.manager [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1170.305052] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1170.306014] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1328cf6a-b21a-4302-b405-3db0660b2e9f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.314362] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1170.314648] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-234fe680-f223-4da3-86f3-dc817a3cbcc7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.321252] env[65522]: DEBUG oslo_vmware.api [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1170.321252] env[65522]: value = "task-5114940" [ 1170.321252] env[65522]: _type = "Task" [ 1170.321252] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.330763] env[65522]: DEBUG oslo_vmware.api [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114940, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.510217] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c4941560-4730-4ccc-a1be-d40011725fea tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "2eb8e092-255f-4997-8f10-a0cce0f4d503" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.781s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1170.623948] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea386cde-8bc3-4cec-b58c-6e12f9c350fe tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Lock "6d856111-81bd-4e2f-b7fc-e169c3e974a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.180s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1170.630347] env[65522]: DEBUG nova.compute.manager [req-4fd5f4a1-d884-4574-9433-ccef2f6566f7 req-0ac33c95-c781-4641-8214-01fd1ef95643 service nova] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Received event network-vif-plugged-0b80621b-795f-4503-9161-c9799b6339b6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1170.630643] env[65522]: DEBUG oslo_concurrency.lockutils [req-4fd5f4a1-d884-4574-9433-ccef2f6566f7 req-0ac33c95-c781-4641-8214-01fd1ef95643 service nova] Acquiring lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1170.630792] env[65522]: DEBUG oslo_concurrency.lockutils [req-4fd5f4a1-d884-4574-9433-ccef2f6566f7 req-0ac33c95-c781-4641-8214-01fd1ef95643 service nova] Lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1170.631022] env[65522]: DEBUG oslo_concurrency.lockutils [req-4fd5f4a1-d884-4574-9433-ccef2f6566f7 req-0ac33c95-c781-4641-8214-01fd1ef95643 service nova] Lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1170.631223] env[65522]: DEBUG nova.compute.manager [req-4fd5f4a1-d884-4574-9433-ccef2f6566f7 req-0ac33c95-c781-4641-8214-01fd1ef95643 service nova] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] No waiting events found dispatching network-vif-plugged-0b80621b-795f-4503-9161-c9799b6339b6 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1170.631362] env[65522]: WARNING nova.compute.manager [req-4fd5f4a1-d884-4574-9433-ccef2f6566f7 req-0ac33c95-c781-4641-8214-01fd1ef95643 service nova] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Received unexpected event network-vif-plugged-0b80621b-795f-4503-9161-c9799b6339b6 for instance with vm_state building and task_state spawning. [ 1170.648534] env[65522]: DEBUG nova.network.neutron [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Successfully updated port: 0b80621b-795f-4503-9161-c9799b6339b6 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1170.700554] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ebd804-fc31-4a87-baea-adfe36b5861f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.708821] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eeb7028-0bbc-49a4-b37d-dce7f2d7bfbe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.741152] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1be1157-d18c-4e94-9a6f-eb5f9bc0d5e9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.749289] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da53a4f0-89b5-4e59-9b46-db028c65ff93 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.764028] env[65522]: DEBUG nova.compute.provider_tree [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1170.833807] env[65522]: DEBUG oslo_vmware.api [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114940, 'name': PowerOffVM_Task, 'duration_secs': 0.341097} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.834167] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1170.834372] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1170.834773] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2a55b69-e8b1-412e-aa68-e7b1c145119e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.909344] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1170.909566] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1170.909742] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleting the datastore file [datastore1] 9f342b89-bde2-4c35-ae42-cfe1e6973b74 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1170.910037] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c0309e9-6c52-42a1-8c90-13770f1a11a6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.917814] env[65522]: DEBUG oslo_vmware.api [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1170.917814] env[65522]: value = "task-5114942" [ 1170.917814] env[65522]: _type = "Task" [ 1170.917814] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.926756] env[65522]: DEBUG oslo_vmware.api [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114942, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.151213] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "refresh_cache-aebb7082-f88c-42f3-9d9f-93cb2a6f545b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.151481] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "refresh_cache-aebb7082-f88c-42f3-9d9f-93cb2a6f545b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1171.151616] env[65522]: DEBUG nova.network.neutron [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1171.267210] env[65522]: DEBUG nova.scheduler.client.report [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1171.429071] env[65522]: DEBUG oslo_vmware.api [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5114942, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.30897} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.429071] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1171.429071] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1171.429071] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1171.429313] env[65522]: INFO nova.compute.manager [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1171.429587] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1171.429841] env[65522]: DEBUG nova.compute.manager [-] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1171.429987] env[65522]: DEBUG nova.network.neutron [-] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1171.430303] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1171.430870] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1171.431179] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1171.452277] env[65522]: INFO nova.compute.manager [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Rebuilding instance [ 1171.469641] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1171.489061] env[65522]: DEBUG nova.compute.manager [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1171.489955] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778a299d-08a4-450d-8342-7487369f3e03 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.570177] env[65522]: WARNING neutronclient.v2_0.client [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1171.607606] env[65522]: WARNING neutronclient.v2_0.client [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1171.608078] env[65522]: WARNING neutronclient.v2_0.client [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1171.651617] env[65522]: DEBUG nova.network.neutron [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Port a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac binding to destination host cpu-1 is already ACTIVE {{(pid=65522) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 1171.651897] env[65522]: DEBUG oslo_concurrency.lockutils [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.652057] env[65522]: DEBUG oslo_concurrency.lockutils [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1171.652222] env[65522]: DEBUG nova.network.neutron [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1171.655073] env[65522]: WARNING openstack [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1171.655073] env[65522]: WARNING openstack [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1171.708131] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "4aca4e87-b923-49e4-88d8-7ff51bcbe993" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1171.708475] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "4aca4e87-b923-49e4-88d8-7ff51bcbe993" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.708746] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "4aca4e87-b923-49e4-88d8-7ff51bcbe993-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1171.708990] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "4aca4e87-b923-49e4-88d8-7ff51bcbe993-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.709240] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "4aca4e87-b923-49e4-88d8-7ff51bcbe993-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1171.712198] env[65522]: INFO nova.compute.manager [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Terminating instance [ 1171.714467] env[65522]: DEBUG nova.network.neutron [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1171.757128] env[65522]: WARNING openstack [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1171.757589] env[65522]: WARNING openstack [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1171.773369] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.792s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1171.775967] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.880s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.776373] env[65522]: DEBUG nova.objects.instance [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lazy-loading 'resources' on Instance uuid b93f2d74-a24b-4f2c-aefd-16b093107ec4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1171.795275] env[65522]: INFO nova.scheduler.client.report [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Deleted allocations for instance 59cc1b4f-b451-41d4-a071-b1ec0751bf38 [ 1171.829361] env[65522]: WARNING neutronclient.v2_0.client [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1171.830319] env[65522]: WARNING openstack [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1171.830574] env[65522]: WARNING openstack [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1171.929699] env[65522]: DEBUG nova.network.neutron [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Updating instance_info_cache with network_info: [{"id": "0b80621b-795f-4503-9161-c9799b6339b6", "address": "fa:16:3e:e7:ac:80", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b80621b-79", "ovs_interfaceid": "0b80621b-795f-4503-9161-c9799b6339b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1171.975688] env[65522]: DEBUG nova.compute.manager [req-cb7cee59-3517-4911-8268-0f00aca00751 req-e013dea0-9d6f-407a-b5ac-8303288e740a service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Received event network-vif-deleted-ed6e9224-69f8-46fd-a71f-27d8bc6eaec3 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1171.975896] env[65522]: INFO nova.compute.manager [req-cb7cee59-3517-4911-8268-0f00aca00751 req-e013dea0-9d6f-407a-b5ac-8303288e740a service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Neutron deleted interface ed6e9224-69f8-46fd-a71f-27d8bc6eaec3; detaching it from the instance and deleting it from the info cache [ 1171.976092] env[65522]: DEBUG nova.network.neutron [req-cb7cee59-3517-4911-8268-0f00aca00751 req-e013dea0-9d6f-407a-b5ac-8303288e740a service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1172.154622] env[65522]: WARNING neutronclient.v2_0.client [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1172.155788] env[65522]: WARNING openstack [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1172.156248] env[65522]: WARNING openstack [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1172.222963] env[65522]: DEBUG nova.compute.manager [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1172.223371] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1172.224413] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b46fa5c7-92f3-4260-bc8a-f29d0173b538 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.233970] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1172.234467] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4397fe90-8e1c-4f1e-a5e7-14aad1c19a25 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.249909] env[65522]: DEBUG oslo_vmware.api [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1172.249909] env[65522]: value = "task-5114943" [ 1172.249909] env[65522]: _type = "Task" [ 1172.249909] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.261136] env[65522]: DEBUG oslo_vmware.api [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114943, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.281511] env[65522]: WARNING openstack [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1172.284078] env[65522]: WARNING openstack [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1172.304081] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0255dd0a-ed2d-40de-bdc8-47ec4cbfd2f2 tempest-ServerMetadataTestJSON-383955296 tempest-ServerMetadataTestJSON-383955296-project-member] Lock "59cc1b4f-b451-41d4-a071-b1ec0751bf38" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.370s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.378611] env[65522]: WARNING neutronclient.v2_0.client [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1172.379427] env[65522]: WARNING openstack [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1172.379801] env[65522]: WARNING openstack [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1172.433061] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "refresh_cache-aebb7082-f88c-42f3-9d9f-93cb2a6f545b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1172.433548] env[65522]: DEBUG nova.compute.manager [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Instance network_info: |[{"id": "0b80621b-795f-4503-9161-c9799b6339b6", "address": "fa:16:3e:e7:ac:80", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b80621b-79", "ovs_interfaceid": "0b80621b-795f-4503-9161-c9799b6339b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1172.436916] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:ac:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56136ef6-99d7-4562-9a9f-d66fec951c5c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0b80621b-795f-4503-9161-c9799b6339b6', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1172.445881] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1172.447334] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1172.447334] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8578a2be-f926-4aa3-b4ee-e08b9cbe5039 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.464385] env[65522]: DEBUG nova.network.neutron [-] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1172.477240] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1172.477240] env[65522]: value = "task-5114944" [ 1172.477240] env[65522]: _type = "Task" [ 1172.477240] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.483478] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6d85b72f-192a-42ae-bd17-0769e9c67434 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.490994] env[65522]: DEBUG nova.network.neutron [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance_info_cache with network_info: [{"id": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "address": "fa:16:3e:5b:f2:d8", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7fc03fa-06", "ovs_interfaceid": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1172.499268] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114944, 'name': CreateVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.504737] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-344ec0a0-af9f-4993-91db-f540639e4eaf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.521449] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1172.522188] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1cc03b81-5576-4b96-b43c-c6651f6e738f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.531930] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1172.531930] env[65522]: value = "task-5114945" [ 1172.531930] env[65522]: _type = "Task" [ 1172.531930] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.548649] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97634b0-ac3f-4dad-ae38-454f61e186e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.553095] env[65522]: DEBUG nova.compute.manager [req-cb7cee59-3517-4911-8268-0f00aca00751 req-e013dea0-9d6f-407a-b5ac-8303288e740a service nova] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Detach interface failed, port_id=ed6e9224-69f8-46fd-a71f-27d8bc6eaec3, reason: Instance 9f342b89-bde2-4c35-ae42-cfe1e6973b74 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1172.562784] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f114b6-76e0-4d1a-b51d-9c86b674eb15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.566864] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114945, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.604246] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d656f6c7-bb86-410f-8c13-a479290a7aa6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.615442] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ddf2925-192c-4ecc-83ef-f811c5be0f60 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.631712] env[65522]: DEBUG nova.compute.provider_tree [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1172.672502] env[65522]: DEBUG nova.compute.manager [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Received event network-changed-0b80621b-795f-4503-9161-c9799b6339b6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1172.672502] env[65522]: DEBUG nova.compute.manager [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Refreshing instance network info cache due to event network-changed-0b80621b-795f-4503-9161-c9799b6339b6. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1172.672502] env[65522]: DEBUG oslo_concurrency.lockutils [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] Acquiring lock "refresh_cache-aebb7082-f88c-42f3-9d9f-93cb2a6f545b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.672502] env[65522]: DEBUG oslo_concurrency.lockutils [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] Acquired lock "refresh_cache-aebb7082-f88c-42f3-9d9f-93cb2a6f545b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1172.672641] env[65522]: DEBUG nova.network.neutron [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Refreshing network info cache for port 0b80621b-795f-4503-9161-c9799b6339b6 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1172.761872] env[65522]: DEBUG oslo_vmware.api [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114943, 'name': PowerOffVM_Task, 'duration_secs': 0.380863} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.762189] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1172.762356] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1172.762631] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aca7fe0d-8407-49f5-9016-b7433d918651 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.838575] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1172.838948] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1172.839223] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleting the datastore file [datastore2] 4aca4e87-b923-49e4-88d8-7ff51bcbe993 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1172.839583] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf13db3b-b7ee-4a8a-9673-29ec7dd1fb77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.850218] env[65522]: DEBUG oslo_vmware.api [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for the task: (returnval){ [ 1172.850218] env[65522]: value = "task-5114947" [ 1172.850218] env[65522]: _type = "Task" [ 1172.850218] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.862451] env[65522]: DEBUG oslo_vmware.api [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114947, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.970738] env[65522]: INFO nova.compute.manager [-] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Took 1.54 seconds to deallocate network for instance. [ 1172.988628] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114944, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.994506] env[65522]: DEBUG oslo_concurrency.lockutils [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1173.044429] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114945, 'name': PowerOffVM_Task, 'duration_secs': 0.167247} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.044429] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1173.044429] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1173.045416] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4c4acc-4fde-4a2c-ab33-c3fa66be6bb3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.054538] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1173.054818] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05eee589-a39b-49b2-8bc3-32d9e0901d06 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.082549] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1173.082768] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1173.083053] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Deleting the datastore file [datastore1] 6d856111-81bd-4e2f-b7fc-e169c3e974a6 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1173.083368] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e7b8ec54-025c-4344-921c-be938c7d1c1e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.092026] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1173.092026] env[65522]: value = "task-5114949" [ 1173.092026] env[65522]: _type = "Task" [ 1173.092026] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.101873] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114949, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.135252] env[65522]: DEBUG nova.scheduler.client.report [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1173.176660] env[65522]: WARNING neutronclient.v2_0.client [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1173.177871] env[65522]: WARNING openstack [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1173.177871] env[65522]: WARNING openstack [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1173.366655] env[65522]: DEBUG oslo_vmware.api [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Task: {'id': task-5114947, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.320619} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.366927] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1173.367190] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1173.367353] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1173.367525] env[65522]: INFO nova.compute.manager [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1173.367772] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1173.367981] env[65522]: DEBUG nova.compute.manager [-] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1173.368118] env[65522]: DEBUG nova.network.neutron [-] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1173.368634] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1173.368871] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1173.369159] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1173.378440] env[65522]: WARNING openstack [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1173.378846] env[65522]: WARNING openstack [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1173.455130] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1173.478220] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1173.502790] env[65522]: DEBUG nova.compute.manager [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65522) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:924}} [ 1173.503217] env[65522]: DEBUG oslo_concurrency.lockutils [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1173.503646] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114944, 'name': CreateVM_Task, 'duration_secs': 0.538529} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.503841] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1173.505437] env[65522]: WARNING neutronclient.v2_0.client [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1173.505437] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1173.505437] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1173.506299] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1173.506299] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e2aa2e0-8b91-4d6f-9389-fa08fbc37541 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.515639] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1173.515639] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f2d067-7890-9c75-d9ce-e56cf083ecb5" [ 1173.515639] env[65522]: _type = "Task" [ 1173.515639] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.528980] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f2d067-7890-9c75-d9ce-e56cf083ecb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.604797] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114949, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234299} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.605313] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1173.605657] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1173.605739] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1173.641901] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.866s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1173.647838] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.378s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1173.648154] env[65522]: DEBUG nova.objects.instance [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lazy-loading 'resources' on Instance uuid af9e93f9-3e6f-4267-b7d9-1961af38c84f {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1173.675602] env[65522]: INFO nova.scheduler.client.report [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Deleted allocations for instance b93f2d74-a24b-4f2c-aefd-16b093107ec4 [ 1173.873290] env[65522]: WARNING neutronclient.v2_0.client [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1173.874164] env[65522]: WARNING openstack [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1173.874575] env[65522]: WARNING openstack [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1174.003957] env[65522]: DEBUG nova.network.neutron [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Updated VIF entry in instance network info cache for port 0b80621b-795f-4503-9161-c9799b6339b6. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1174.004253] env[65522]: DEBUG nova.network.neutron [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Updating instance_info_cache with network_info: [{"id": "0b80621b-795f-4503-9161-c9799b6339b6", "address": "fa:16:3e:e7:ac:80", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b80621b-79", "ovs_interfaceid": "0b80621b-795f-4503-9161-c9799b6339b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1174.033175] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f2d067-7890-9c75-d9ce-e56cf083ecb5, 'name': SearchDatastore_Task, 'duration_secs': 0.014256} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.033649] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1174.033978] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1174.034347] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1174.034541] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1174.034863] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1174.035111] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-28630883-653b-4798-9c56-b2f334c25dab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.048016] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1174.048634] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1174.049247] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7434b589-f0ab-427d-b3cd-5bacc52a4bfc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.056275] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1174.056275] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52abe967-341c-c6f6-1dc9-482ec7b15b8b" [ 1174.056275] env[65522]: _type = "Task" [ 1174.056275] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.065768] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52abe967-341c-c6f6-1dc9-482ec7b15b8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.188325] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d89ee30a-0848-4dc8-aa8e-acf5d928a1bc tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "b93f2d74-a24b-4f2c-aefd-16b093107ec4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.785s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1174.316894] env[65522]: DEBUG nova.network.neutron [-] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1174.332551] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815a6b20-e343-4882-82e0-07cb9857c4fd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.339910] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693f2d81-5c48-4611-85a2-f8001f7470dc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.377613] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68315035-2331-4cc8-925e-032003ae9073 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.387210] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b85afcf-1715-4787-acca-e320f3fd9060 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.404652] env[65522]: DEBUG nova.compute.provider_tree [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1174.510960] env[65522]: DEBUG oslo_concurrency.lockutils [req-0553db2d-f114-464b-a84f-817bad9d17d7 req-5544192f-4bf8-4ec7-90b2-72095ba76b83 service nova] Releasing lock "refresh_cache-aebb7082-f88c-42f3-9d9f-93cb2a6f545b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1174.569183] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52abe967-341c-c6f6-1dc9-482ec7b15b8b, 'name': SearchDatastore_Task, 'duration_secs': 0.013988} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.570582] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a77b3e1-634b-4ba4-b690-98438fe56086 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.577369] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1174.577369] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5289adde-cf43-75e0-bfa7-8cf2cde6ba23" [ 1174.577369] env[65522]: _type = "Task" [ 1174.577369] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.588328] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5289adde-cf43-75e0-bfa7-8cf2cde6ba23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.654882] env[65522]: DEBUG nova.virt.hardware [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1174.654882] env[65522]: DEBUG nova.virt.hardware [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1174.654986] env[65522]: DEBUG nova.virt.hardware [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1174.657376] env[65522]: DEBUG nova.virt.hardware [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1174.657376] env[65522]: DEBUG nova.virt.hardware [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1174.657376] env[65522]: DEBUG nova.virt.hardware [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1174.657376] env[65522]: DEBUG nova.virt.hardware [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1174.657607] env[65522]: DEBUG nova.virt.hardware [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1174.657847] env[65522]: DEBUG nova.virt.hardware [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1174.658501] env[65522]: DEBUG nova.virt.hardware [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1174.658746] env[65522]: DEBUG nova.virt.hardware [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1174.659668] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66546295-3aad-4e85-8d23-43a5275b0826 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.672721] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc1da69-900a-454e-b085-4db0f962f7b4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.689663] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1174.698376] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1174.698613] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1174.698857] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0e1a5b4-e2e1-4700-a81f-cb84b2b857c8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.715665] env[65522]: DEBUG nova.compute.manager [req-576a7f95-a1b4-4c19-8c5c-83d30ca97c8d req-b74b319e-0444-4955-84a6-e95ed0479918 service nova] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Received event network-vif-deleted-1cc60bc4-0fc3-4b43-8ebf-11ac6fca59a2 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1174.724311] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1174.724311] env[65522]: value = "task-5114950" [ 1174.724311] env[65522]: _type = "Task" [ 1174.724311] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.734186] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114950, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.820306] env[65522]: INFO nova.compute.manager [-] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Took 1.45 seconds to deallocate network for instance. [ 1174.908020] env[65522]: DEBUG nova.scheduler.client.report [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1175.089584] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5289adde-cf43-75e0-bfa7-8cf2cde6ba23, 'name': SearchDatastore_Task, 'duration_secs': 0.0115} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.089967] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1175.090659] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] aebb7082-f88c-42f3-9d9f-93cb2a6f545b/aebb7082-f88c-42f3-9d9f-93cb2a6f545b.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1175.090780] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-92b8d4dc-a04c-43b2-8dad-d0208c169cc1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.099179] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1175.099179] env[65522]: value = "task-5114951" [ 1175.099179] env[65522]: _type = "Task" [ 1175.099179] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.110865] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114951, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.235173] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114950, 'name': CreateVM_Task, 'duration_secs': 0.344466} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.235612] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1175.235816] env[65522]: DEBUG oslo_concurrency.lockutils [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.236030] env[65522]: DEBUG oslo_concurrency.lockutils [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1175.236307] env[65522]: DEBUG oslo_concurrency.lockutils [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1175.236587] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f4729d4-5e3b-4292-9c19-87ca3cc41e10 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.242897] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1175.242897] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fdb5e3-89bf-bf7a-7022-e78ca82e3a33" [ 1175.242897] env[65522]: _type = "Task" [ 1175.242897] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.251973] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fdb5e3-89bf-bf7a-7022-e78ca82e3a33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.328257] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1175.413546] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.765s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1175.417050] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.716s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1175.418310] env[65522]: INFO nova.compute.claims [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1175.438194] env[65522]: INFO nova.scheduler.client.report [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleted allocations for instance af9e93f9-3e6f-4267-b7d9-1961af38c84f [ 1175.611962] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114951, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.754568] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52fdb5e3-89bf-bf7a-7022-e78ca82e3a33, 'name': SearchDatastore_Task, 'duration_secs': 0.02661} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.754901] env[65522]: DEBUG oslo_concurrency.lockutils [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1175.755148] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1175.755425] env[65522]: DEBUG oslo_concurrency.lockutils [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.755603] env[65522]: DEBUG oslo_concurrency.lockutils [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1175.755795] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1175.756087] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-55b31647-9007-4d84-aca9-f08a3a6ae3a9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.765933] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1175.766143] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1175.766906] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99b8e349-4dfe-4f20-89d2-e812672c183f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.772368] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1175.772368] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527de1a8-c78d-7815-0878-09d83fea3838" [ 1175.772368] env[65522]: _type = "Task" [ 1175.772368] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.781143] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527de1a8-c78d-7815-0878-09d83fea3838, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.948728] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bded13aa-e856-4d27-b278-4ceffbe8e6c7 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "af9e93f9-3e6f-4267-b7d9-1961af38c84f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.699s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1176.110536] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114951, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515884} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.110794] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] aebb7082-f88c-42f3-9d9f-93cb2a6f545b/aebb7082-f88c-42f3-9d9f-93cb2a6f545b.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1176.111012] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1176.111287] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ba457767-f29e-42e1-9f22-114fefb31bd9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.119551] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1176.119551] env[65522]: value = "task-5114952" [ 1176.119551] env[65522]: _type = "Task" [ 1176.119551] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.128254] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114952, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.284294] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527de1a8-c78d-7815-0878-09d83fea3838, 'name': SearchDatastore_Task, 'duration_secs': 0.010597} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.285148] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c06ba49-3a46-4507-9f9c-a1ca6fc6f16a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.291461] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1176.291461] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528fab52-48d5-5807-6444-97b37a66273f" [ 1176.291461] env[65522]: _type = "Task" [ 1176.291461] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.301437] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528fab52-48d5-5807-6444-97b37a66273f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.614791] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab0d92b-8763-4d69-b0c8-41d4d26fcf97 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.626152] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c234fc95-27c1-4c3b-aa3a-6354cee69ae8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.638210] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114952, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.137536} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.663813] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1176.664984] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519b5d4a-39e5-4275-8f8c-a8535ea1ef50 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.668146] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a104069-83f3-46dd-bd84-e53c0c8cdcb7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.693709] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] aebb7082-f88c-42f3-9d9f-93cb2a6f545b/aebb7082-f88c-42f3-9d9f-93cb2a6f545b.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1176.695536] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55bb3f7c-e78d-40ec-89a1-42fa795b1d51 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.712051] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3dd5ee3-eccd-4403-ad76-cf2d3e9279e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.728999] env[65522]: DEBUG nova.compute.provider_tree [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1176.733045] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1176.733045] env[65522]: value = "task-5114953" [ 1176.733045] env[65522]: _type = "Task" [ 1176.733045] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.743284] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114953, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.807029] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528fab52-48d5-5807-6444-97b37a66273f, 'name': SearchDatastore_Task, 'duration_secs': 0.011334} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.807029] env[65522]: DEBUG oslo_concurrency.lockutils [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1176.807217] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 6d856111-81bd-4e2f-b7fc-e169c3e974a6/6d856111-81bd-4e2f-b7fc-e169c3e974a6.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1176.807334] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b52bacb-8798-4ed8-90f4-4227f79bc985 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.814744] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1176.814744] env[65522]: value = "task-5114954" [ 1176.814744] env[65522]: _type = "Task" [ 1176.814744] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.826019] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114954, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.232341] env[65522]: DEBUG nova.scheduler.client.report [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1177.256797] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114953, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.295694] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1177.298105] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1177.329189] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114954, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.743869] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.327s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1177.744369] env[65522]: DEBUG nova.compute.manager [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1177.750738] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 7.688s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1177.752603] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114953, 'name': ReconfigVM_Task, 'duration_secs': 0.840316} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.753639] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Reconfigured VM instance instance-0000006f to attach disk [datastore1] aebb7082-f88c-42f3-9d9f-93cb2a6f545b/aebb7082-f88c-42f3-9d9f-93cb2a6f545b.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1177.753795] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c186d7c6-aa55-43f7-bab3-e65cb10e89e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.762336] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1177.762336] env[65522]: value = "task-5114955" [ 1177.762336] env[65522]: _type = "Task" [ 1177.762336] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.773860] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114955, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.800179] env[65522]: DEBUG nova.compute.manager [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1177.828177] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114954, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.261128] env[65522]: DEBUG nova.compute.utils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1178.262067] env[65522]: DEBUG nova.compute.manager [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1178.262221] env[65522]: DEBUG nova.network.neutron [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1178.262522] env[65522]: WARNING neutronclient.v2_0.client [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1178.262823] env[65522]: WARNING neutronclient.v2_0.client [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1178.263544] env[65522]: WARNING openstack [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1178.263845] env[65522]: WARNING openstack [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1178.282858] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114955, 'name': Rename_Task, 'duration_secs': 0.317665} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.283110] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1178.283382] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-89243490-1e27-4a0b-b5d6-31b495e3266c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.292036] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1178.292036] env[65522]: value = "task-5114956" [ 1178.292036] env[65522]: _type = "Task" [ 1178.292036] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.301459] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114956, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.312189] env[65522]: DEBUG nova.policy [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '48bde7741f2d4bfabdb803d10cf79599', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a80f979f3dc0477e9462b47f7aa87f14', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1178.329112] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114954, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.330474] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1178.695265] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "874b6a07-ebba-4f75-937e-2281af1598b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1178.698876] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "874b6a07-ebba-4f75-937e-2281af1598b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1178.700494] env[65522]: DEBUG nova.network.neutron [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Successfully created port: c8c5ef6a-c86a-4a39-a2cc-918754e4e78a {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1178.765027] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Applying migration context for instance cb62df50-2f86-469a-b2ee-fd3754d61c83 as it has an incoming, in-progress migration 76ba5bb4-2a61-440b-a450-56d64e13a6a0. Migration status is reverting {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1178.765027] env[65522]: INFO nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating resource usage from migration 76ba5bb4-2a61-440b-a450-56d64e13a6a0 [ 1178.774619] env[65522]: DEBUG nova.compute.manager [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1178.788560] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7617189c-a902-42e7-8165-0e7c4a1de06d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1178.788560] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 4aca4e87-b923-49e4-88d8-7ff51bcbe993 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1178.788560] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 2981b635-43c8-4bd6-9991-e6af0be82f3c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1178.788560] env[65522]: WARNING nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 9f342b89-bde2-4c35-ae42-cfe1e6973b74 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1178.788560] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 3b82f9ce-ea59-430d-adb6-918cc6cc48a4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1178.788560] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Migration 76ba5bb4-2a61-440b-a450-56d64e13a6a0 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1178.788560] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance cb62df50-2f86-469a-b2ee-fd3754d61c83 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1178.788560] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 6d856111-81bd-4e2f-b7fc-e169c3e974a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1178.788560] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance aebb7082-f88c-42f3-9d9f-93cb2a6f545b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1178.788560] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 3987d71f-47ab-4dd4-8426-95d78503effa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1178.803207] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114956, 'name': PowerOnVM_Task} progress is 81%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.829817] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114954, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.532378} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.830129] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 6d856111-81bd-4e2f-b7fc-e169c3e974a6/6d856111-81bd-4e2f-b7fc-e169c3e974a6.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1178.830352] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1178.830886] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0e0b3a19-979a-4e22-a767-f3c6b5c5d0f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.840693] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1178.840693] env[65522]: value = "task-5114957" [ 1178.840693] env[65522]: _type = "Task" [ 1178.840693] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.851850] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114957, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.207382] env[65522]: DEBUG nova.compute.manager [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1179.290913] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance ae3455d8-2f38-42ad-b16d-d98aef92b2a8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1179.308018] env[65522]: DEBUG oslo_vmware.api [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5114956, 'name': PowerOnVM_Task, 'duration_secs': 0.95038} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.308018] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1179.308341] env[65522]: INFO nova.compute.manager [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Took 9.10 seconds to spawn the instance on the hypervisor. [ 1179.308379] env[65522]: DEBUG nova.compute.manager [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1179.309188] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e60583fa-19be-4919-823c-62ba48704877 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.352522] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114957, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078765} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.353404] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1179.353955] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf0473e-3a44-41fd-b966-86c1eb54eb0b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.379781] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 6d856111-81bd-4e2f-b7fc-e169c3e974a6/6d856111-81bd-4e2f-b7fc-e169c3e974a6.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1179.380169] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f4b56f4-ffc2-4549-bef7-f964bcb23b58 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.402571] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1179.402571] env[65522]: value = "task-5114958" [ 1179.402571] env[65522]: _type = "Task" [ 1179.402571] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.414839] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114958, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.731206] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1179.791194] env[65522]: DEBUG nova.compute.manager [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1179.800066] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 874b6a07-ebba-4f75-937e-2281af1598b8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1179.800321] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1179.800480] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=100GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '7', 'num_vm_active': '4', 'num_task_None': '4', 'num_os_type_None': '7', 'num_proj_91597233ae9c44c094f4c32d90332fa6': '2', 'io_workload': '2', 'num_vm_resized': '1', 'num_task_resize_reverting': '1', 'num_proj_fa11b46d9fe144f391233e6eb9c819d7': '1', 'num_proj_fdba2dbc660145b7a98af8bd3f77a035': '1', 'num_proj_1546bbdca07c443d84abea1155cfde37': '1', 'num_task_rebuild_spawning': '1', 'num_proj_ad195a7fa9cc4c2e8cad7ddcc5f0f62a': '1', 'num_vm_building': '2', 'num_task_spawning': '1', 'num_proj_a80f979f3dc0477e9462b47f7aa87f14': '1'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1179.811951] env[65522]: DEBUG nova.virt.hardware [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1179.812204] env[65522]: DEBUG nova.virt.hardware [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1179.812359] env[65522]: DEBUG nova.virt.hardware [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1179.812572] env[65522]: DEBUG nova.virt.hardware [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1179.812726] env[65522]: DEBUG nova.virt.hardware [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1179.812871] env[65522]: DEBUG nova.virt.hardware [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1179.813110] env[65522]: DEBUG nova.virt.hardware [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1179.813314] env[65522]: DEBUG nova.virt.hardware [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1179.813458] env[65522]: DEBUG nova.virt.hardware [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1179.813619] env[65522]: DEBUG nova.virt.hardware [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1179.813788] env[65522]: DEBUG nova.virt.hardware [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1179.814949] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3434304-c1b2-4bfa-9f24-148f1d0c8829 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.831678] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd2e5cee-8d0d-4801-a3d4-233cf8376eb5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.840518] env[65522]: INFO nova.compute.manager [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Took 22.40 seconds to build instance. [ 1179.913708] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114958, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.969882] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8757725-148f-4f37-b65d-5ed8f30d9e10 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.978460] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b68fc8d-210d-43b3-85bd-83e63496cbca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.011335] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22bf29c-74fb-467f-b89e-815420f8aead {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.019817] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40f81c2-90b6-4c96-b241-d2cf96ec303e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.038317] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1180.148443] env[65522]: DEBUG nova.compute.manager [req-681108a5-22de-4216-ad8d-fc343ec531e4 req-bbe9b934-88d7-4723-9eb5-2dfe972e1930 service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Received event network-vif-plugged-c8c5ef6a-c86a-4a39-a2cc-918754e4e78a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1180.148544] env[65522]: DEBUG oslo_concurrency.lockutils [req-681108a5-22de-4216-ad8d-fc343ec531e4 req-bbe9b934-88d7-4723-9eb5-2dfe972e1930 service nova] Acquiring lock "3987d71f-47ab-4dd4-8426-95d78503effa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1180.148687] env[65522]: DEBUG oslo_concurrency.lockutils [req-681108a5-22de-4216-ad8d-fc343ec531e4 req-bbe9b934-88d7-4723-9eb5-2dfe972e1930 service nova] Lock "3987d71f-47ab-4dd4-8426-95d78503effa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1180.148849] env[65522]: DEBUG oslo_concurrency.lockutils [req-681108a5-22de-4216-ad8d-fc343ec531e4 req-bbe9b934-88d7-4723-9eb5-2dfe972e1930 service nova] Lock "3987d71f-47ab-4dd4-8426-95d78503effa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1180.149178] env[65522]: DEBUG nova.compute.manager [req-681108a5-22de-4216-ad8d-fc343ec531e4 req-bbe9b934-88d7-4723-9eb5-2dfe972e1930 service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] No waiting events found dispatching network-vif-plugged-c8c5ef6a-c86a-4a39-a2cc-918754e4e78a {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1180.149399] env[65522]: WARNING nova.compute.manager [req-681108a5-22de-4216-ad8d-fc343ec531e4 req-bbe9b934-88d7-4723-9eb5-2dfe972e1930 service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Received unexpected event network-vif-plugged-c8c5ef6a-c86a-4a39-a2cc-918754e4e78a for instance with vm_state building and task_state spawning. [ 1180.249618] env[65522]: DEBUG nova.network.neutron [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Successfully updated port: c8c5ef6a-c86a-4a39-a2cc-918754e4e78a {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1180.343583] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6317dc3d-1f78-4501-88bd-4a328d69ca18 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.914s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1180.416469] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114958, 'name': ReconfigVM_Task, 'duration_secs': 0.655147} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.416755] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 6d856111-81bd-4e2f-b7fc-e169c3e974a6/6d856111-81bd-4e2f-b7fc-e169c3e974a6.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1180.417390] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d1b5d3a0-7f5d-44e7-ace4-b5a91c0bc6b3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.423787] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1180.423787] env[65522]: value = "task-5114959" [ 1180.423787] env[65522]: _type = "Task" [ 1180.423787] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.432061] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114959, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.542100] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1180.726502] env[65522]: DEBUG nova.compute.manager [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Received event network-changed-0b80621b-795f-4503-9161-c9799b6339b6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1180.726695] env[65522]: DEBUG nova.compute.manager [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Refreshing instance network info cache due to event network-changed-0b80621b-795f-4503-9161-c9799b6339b6. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1180.726915] env[65522]: DEBUG oslo_concurrency.lockutils [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] Acquiring lock "refresh_cache-aebb7082-f88c-42f3-9d9f-93cb2a6f545b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1180.727062] env[65522]: DEBUG oslo_concurrency.lockutils [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] Acquired lock "refresh_cache-aebb7082-f88c-42f3-9d9f-93cb2a6f545b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1180.727216] env[65522]: DEBUG nova.network.neutron [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Refreshing network info cache for port 0b80621b-795f-4503-9161-c9799b6339b6 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1180.757531] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "refresh_cache-3987d71f-47ab-4dd4-8426-95d78503effa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1180.757531] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquired lock "refresh_cache-3987d71f-47ab-4dd4-8426-95d78503effa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1180.757531] env[65522]: DEBUG nova.network.neutron [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1180.933814] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114959, 'name': Rename_Task, 'duration_secs': 0.22079} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.934134] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1180.934373] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ff39bf1-cd37-43e0-bac5-92abadbb7007 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.940972] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1180.940972] env[65522]: value = "task-5114960" [ 1180.940972] env[65522]: _type = "Task" [ 1180.940972] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.953051] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114960, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.047349] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1181.047650] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.297s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1181.047873] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.570s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1181.048087] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1181.050428] env[65522]: DEBUG oslo_concurrency.lockutils [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 7.547s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1181.051873] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1181.052040] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Cleaning up deleted instances {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11933}} [ 1181.081743] env[65522]: INFO nova.scheduler.client.report [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleted allocations for instance 9f342b89-bde2-4c35-ae42-cfe1e6973b74 [ 1181.230119] env[65522]: WARNING neutronclient.v2_0.client [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1181.234126] env[65522]: WARNING openstack [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1181.234126] env[65522]: WARNING openstack [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1181.260595] env[65522]: WARNING openstack [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1181.261863] env[65522]: WARNING openstack [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1181.310419] env[65522]: DEBUG nova.network.neutron [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1181.333551] env[65522]: WARNING openstack [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1181.334013] env[65522]: WARNING openstack [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1181.379685] env[65522]: WARNING openstack [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1181.380107] env[65522]: WARNING openstack [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1181.425182] env[65522]: WARNING neutronclient.v2_0.client [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1181.425818] env[65522]: WARNING openstack [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1181.426179] env[65522]: WARNING openstack [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1181.451764] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114960, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.467923] env[65522]: WARNING neutronclient.v2_0.client [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1181.468582] env[65522]: WARNING openstack [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1181.468933] env[65522]: WARNING openstack [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1181.555437] env[65522]: DEBUG nova.objects.instance [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lazy-loading 'migration_context' on Instance uuid cb62df50-2f86-469a-b2ee-fd3754d61c83 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1181.571692] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] There are 63 instances to clean {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11942}} [ 1181.571871] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: af9e93f9-3e6f-4267-b7d9-1961af38c84f] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1181.576502] env[65522]: DEBUG nova.network.neutron [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Updating instance_info_cache with network_info: [{"id": "c8c5ef6a-c86a-4a39-a2cc-918754e4e78a", "address": "fa:16:3e:ee:34:76", "network": {"id": "f49e9b61-6927-4a5b-a12d-f54f408d42b4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1397158171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a80f979f3dc0477e9462b47f7aa87f14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8c5ef6a-c8", "ovs_interfaceid": "c8c5ef6a-c86a-4a39-a2cc-918754e4e78a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1181.589707] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c2bc5b91-4898-48a6-9b55-8e70c527e464 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "9f342b89-bde2-4c35-ae42-cfe1e6973b74" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.790s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1181.595836] env[65522]: DEBUG nova.network.neutron [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Updated VIF entry in instance network info cache for port 0b80621b-795f-4503-9161-c9799b6339b6. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1181.596459] env[65522]: DEBUG nova.network.neutron [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Updating instance_info_cache with network_info: [{"id": "0b80621b-795f-4503-9161-c9799b6339b6", "address": "fa:16:3e:e7:ac:80", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b80621b-79", "ovs_interfaceid": "0b80621b-795f-4503-9161-c9799b6339b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1181.953307] env[65522]: DEBUG oslo_vmware.api [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114960, 'name': PowerOnVM_Task, 'duration_secs': 0.867481} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.953604] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1181.953818] env[65522]: DEBUG nova.compute.manager [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1181.954703] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff016d06-94c2-4535-9ace-57630a313b71 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.083624] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: b93f2d74-a24b-4f2c-aefd-16b093107ec4] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1182.085856] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Releasing lock "refresh_cache-3987d71f-47ab-4dd4-8426-95d78503effa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1182.086211] env[65522]: DEBUG nova.compute.manager [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Instance network_info: |[{"id": "c8c5ef6a-c86a-4a39-a2cc-918754e4e78a", "address": "fa:16:3e:ee:34:76", "network": {"id": "f49e9b61-6927-4a5b-a12d-f54f408d42b4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1397158171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a80f979f3dc0477e9462b47f7aa87f14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8c5ef6a-c8", "ovs_interfaceid": "c8c5ef6a-c86a-4a39-a2cc-918754e4e78a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1182.086692] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:34:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9d39252e-42ef-4252-98d3-62af5a0d109d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c8c5ef6a-c86a-4a39-a2cc-918754e4e78a', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1182.099030] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1182.099030] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1182.099030] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ace0ea01-f12f-43e5-8da2-e68ef1245cff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.116971] env[65522]: DEBUG oslo_concurrency.lockutils [req-911f8c98-5ef6-4a83-ba33-4f8e9efe686d req-0117e7f6-5a92-451c-9297-38632c7b3fa0 service nova] Releasing lock "refresh_cache-aebb7082-f88c-42f3-9d9f-93cb2a6f545b" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1182.125962] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1182.125962] env[65522]: value = "task-5114962" [ 1182.125962] env[65522]: _type = "Task" [ 1182.125962] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.142014] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114962, 'name': CreateVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.243288] env[65522]: DEBUG nova.compute.manager [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Received event network-changed-c8c5ef6a-c86a-4a39-a2cc-918754e4e78a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1182.243288] env[65522]: DEBUG nova.compute.manager [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Refreshing instance network info cache due to event network-changed-c8c5ef6a-c86a-4a39-a2cc-918754e4e78a. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1182.243288] env[65522]: DEBUG oslo_concurrency.lockutils [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] Acquiring lock "refresh_cache-3987d71f-47ab-4dd4-8426-95d78503effa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.243288] env[65522]: DEBUG oslo_concurrency.lockutils [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] Acquired lock "refresh_cache-3987d71f-47ab-4dd4-8426-95d78503effa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1182.243437] env[65522]: DEBUG nova.network.neutron [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Refreshing network info cache for port c8c5ef6a-c86a-4a39-a2cc-918754e4e78a {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1182.266928] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f960daae-2556-4e59-95d2-dbc077730cc1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.276334] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f780554-3b57-4d26-ad24-189ae55305ce {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.311745] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c35522fc-88f2-475d-a408-af355329fa38 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.321403] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13bac5ae-bae1-4daf-92a0-ea79906c340e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.339249] env[65522]: DEBUG nova.compute.provider_tree [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1182.474498] env[65522]: DEBUG oslo_concurrency.lockutils [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.597525] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 59cc1b4f-b451-41d4-a071-b1ec0751bf38] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1182.639073] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114962, 'name': CreateVM_Task, 'duration_secs': 0.401951} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.639073] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1182.639073] env[65522]: WARNING neutronclient.v2_0.client [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1182.639277] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.639415] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1182.639919] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1182.640210] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-244d365d-abeb-43d7-9fdb-bcc5b1ebbf95 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.645516] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1182.645516] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e4e623-f60d-3f91-e2f7-356a5d81e985" [ 1182.645516] env[65522]: _type = "Task" [ 1182.645516] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.654806] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e4e623-f60d-3f91-e2f7-356a5d81e985, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.746968] env[65522]: WARNING neutronclient.v2_0.client [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1182.747698] env[65522]: WARNING openstack [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1182.748057] env[65522]: WARNING openstack [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1182.843047] env[65522]: DEBUG nova.scheduler.client.report [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1182.859533] env[65522]: WARNING openstack [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1182.859945] env[65522]: WARNING openstack [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1182.922659] env[65522]: WARNING neutronclient.v2_0.client [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1182.923547] env[65522]: WARNING openstack [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1182.923883] env[65522]: WARNING openstack [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1182.940087] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.940316] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1182.962457] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquiring lock "6d856111-81bd-4e2f-b7fc-e169c3e974a6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.962678] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Lock "6d856111-81bd-4e2f-b7fc-e169c3e974a6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1182.962879] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquiring lock "6d856111-81bd-4e2f-b7fc-e169c3e974a6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.963063] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Lock "6d856111-81bd-4e2f-b7fc-e169c3e974a6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1182.963251] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Lock "6d856111-81bd-4e2f-b7fc-e169c3e974a6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1182.965701] env[65522]: INFO nova.compute.manager [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Terminating instance [ 1183.011411] env[65522]: DEBUG nova.network.neutron [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Updated VIF entry in instance network info cache for port c8c5ef6a-c86a-4a39-a2cc-918754e4e78a. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1183.011633] env[65522]: DEBUG nova.network.neutron [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Updating instance_info_cache with network_info: [{"id": "c8c5ef6a-c86a-4a39-a2cc-918754e4e78a", "address": "fa:16:3e:ee:34:76", "network": {"id": "f49e9b61-6927-4a5b-a12d-f54f408d42b4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1397158171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a80f979f3dc0477e9462b47f7aa87f14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8c5ef6a-c8", "ovs_interfaceid": "c8c5ef6a-c86a-4a39-a2cc-918754e4e78a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1183.100927] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 2eb8e092-255f-4997-8f10-a0cce0f4d503] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1183.157500] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e4e623-f60d-3f91-e2f7-356a5d81e985, 'name': SearchDatastore_Task, 'duration_secs': 0.011897} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.158299] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1183.158299] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1183.158299] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.158467] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1183.158545] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1183.158816] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2fc84732-58e3-46d2-8529-39c14891da4d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.168178] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1183.168375] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1183.169118] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1a2c42d-3c45-488a-b78b-a6f7348f5134 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.176699] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1183.176699] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5293109b-d82b-7a9d-47ee-085e468715af" [ 1183.176699] env[65522]: _type = "Task" [ 1183.176699] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.185328] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5293109b-d82b-7a9d-47ee-085e468715af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.442587] env[65522]: DEBUG nova.compute.manager [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1183.469989] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquiring lock "refresh_cache-6d856111-81bd-4e2f-b7fc-e169c3e974a6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.469989] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquired lock "refresh_cache-6d856111-81bd-4e2f-b7fc-e169c3e974a6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1183.470156] env[65522]: DEBUG nova.network.neutron [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1183.514264] env[65522]: DEBUG oslo_concurrency.lockutils [req-751e4eed-dfa7-4241-b6c8-f6b3806bd01c req-2713313b-6404-4386-b0c1-ab8905688f2b service nova] Releasing lock "refresh_cache-3987d71f-47ab-4dd4-8426-95d78503effa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1183.604642] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 95896228-9a70-4f49-bbc9-8f284a666ba9] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1183.688220] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5293109b-d82b-7a9d-47ee-085e468715af, 'name': SearchDatastore_Task, 'duration_secs': 0.012556} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.688656] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8ee0d1b-d0fd-4f34-969e-8e225ebe9096 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.694581] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1183.694581] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528c6baa-7b14-4fe5-9d62-ebfc06fda134" [ 1183.694581] env[65522]: _type = "Task" [ 1183.694581] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.702706] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528c6baa-7b14-4fe5-9d62-ebfc06fda134, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.855225] env[65522]: DEBUG oslo_concurrency.lockutils [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.805s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1183.862070] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.534s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1183.862570] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1183.864632] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.534s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1183.866201] env[65522]: INFO nova.compute.claims [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1183.895944] env[65522]: INFO nova.scheduler.client.report [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Deleted allocations for instance 4aca4e87-b923-49e4-88d8-7ff51bcbe993 [ 1183.963390] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1183.972627] env[65522]: WARNING neutronclient.v2_0.client [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1183.973332] env[65522]: WARNING openstack [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1183.973724] env[65522]: WARNING openstack [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1183.996559] env[65522]: DEBUG nova.network.neutron [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1184.055420] env[65522]: DEBUG nova.network.neutron [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1184.107676] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 3d049de4-4005-4995-a170-5cd6bbf0744c] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1184.206480] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528c6baa-7b14-4fe5-9d62-ebfc06fda134, 'name': SearchDatastore_Task, 'duration_secs': 0.0096} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.206740] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1184.207074] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 3987d71f-47ab-4dd4-8426-95d78503effa/3987d71f-47ab-4dd4-8426-95d78503effa.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1184.207390] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79fb96a4-0427-4780-87d6-7d09307725aa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.215737] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1184.215737] env[65522]: value = "task-5114963" [ 1184.215737] env[65522]: _type = "Task" [ 1184.215737] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.225619] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114963, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.404249] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c5c6fa03-905c-4254-af62-fe63c00e64b7 tempest-ServersTestJSON-546086821 tempest-ServersTestJSON-546086821-project-member] Lock "4aca4e87-b923-49e4-88d8-7ff51bcbe993" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.696s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1184.562341] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Releasing lock "refresh_cache-6d856111-81bd-4e2f-b7fc-e169c3e974a6" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1184.562341] env[65522]: DEBUG nova.compute.manager [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1184.562341] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1184.562341] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89425dad-1eab-4d22-835d-7b32836f8b6a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.580561] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1184.581164] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6053d05f-cc25-46b0-a063-3c94fea0539d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.596227] env[65522]: DEBUG oslo_vmware.api [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1184.596227] env[65522]: value = "task-5114964" [ 1184.596227] env[65522]: _type = "Task" [ 1184.596227] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.609359] env[65522]: DEBUG oslo_vmware.api [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114964, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.611018] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 2b61302b-7ce1-4e8d-b981-8d4376d76c3c] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1185.642065] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 45b4a56e-71b8-42ec-b925-a150e667b223] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1185.643970] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114963, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.644175] env[65522]: WARNING oslo_vmware.common.loopingcall [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] task run outlasted interval by 0.425868 sec [ 1185.660493] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114963, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.551834} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.663556] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 3987d71f-47ab-4dd4-8426-95d78503effa/3987d71f-47ab-4dd4-8426-95d78503effa.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1185.663788] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1185.664102] env[65522]: DEBUG oslo_vmware.api [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114964, 'name': PowerOffVM_Task, 'duration_secs': 0.225663} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.664339] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-87da4eb2-ca56-40c8-95e1-d6dea2a1687e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.666777] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1185.666777] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1185.667174] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dcefc154-257d-4a0b-be40-f25c37c70a79 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.673646] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1185.673646] env[65522]: value = "task-5114966" [ 1185.673646] env[65522]: _type = "Task" [ 1185.673646] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.687910] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114966, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.701952] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1185.702210] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1185.702412] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Deleting the datastore file [datastore1] 6d856111-81bd-4e2f-b7fc-e169c3e974a6 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1185.702741] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5843361a-39c2-4020-864d-b5a36489754a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.717702] env[65522]: DEBUG oslo_vmware.api [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for the task: (returnval){ [ 1185.717702] env[65522]: value = "task-5114968" [ 1185.717702] env[65522]: _type = "Task" [ 1185.717702] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.728174] env[65522]: DEBUG oslo_vmware.api [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114968, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.843935] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ea52ec-551f-4f23-82c1-d83ad5a54828 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.852275] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a68e83f-0d93-4378-859b-f70137182ce3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.893820] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff3259c-a15a-4d00-a526-00e5a1a77b90 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.902520] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-543e5d61-ed81-4ac0-8091-9fb8232cb5d6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.917598] env[65522]: DEBUG nova.compute.provider_tree [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1186.149034] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 07aca0b7-dea4-48b3-a391-58e0e611bfca] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1186.168934] env[65522]: INFO nova.compute.manager [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Swapping old allocation on dict_keys(['9a0d1ac8-ed26-4657-8899-fbceab29db0d']) held by migration 76ba5bb4-2a61-440b-a450-56d64e13a6a0 for instance [ 1186.189846] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114966, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084327} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.190149] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1186.190993] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b0c5a7-d3e2-4518-8017-bff7a6c65fbe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.194662] env[65522]: DEBUG nova.scheduler.client.report [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Overwriting current allocation {'allocations': {'9a0d1ac8-ed26-4657-8899-fbceab29db0d': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 149}}, 'project_id': 'fa11b46d9fe144f391233e6eb9c819d7', 'user_id': '18ceaff46b314c9f9da68a8d0c6f5f2b', 'consumer_generation': 1} on consumer cb62df50-2f86-469a-b2ee-fd3754d61c83 {{(pid=65522) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1186.218373] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] 3987d71f-47ab-4dd4-8426-95d78503effa/3987d71f-47ab-4dd4-8426-95d78503effa.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1186.219190] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb7f99a0-d50c-40bd-93ac-2ae8560fa2b2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.245236] env[65522]: DEBUG oslo_vmware.api [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Task: {'id': task-5114968, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095446} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.246711] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1186.246919] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1186.247109] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1186.247287] env[65522]: INFO nova.compute.manager [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Took 1.69 seconds to destroy the instance on the hypervisor. [ 1186.247540] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1186.247831] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1186.247831] env[65522]: value = "task-5114969" [ 1186.247831] env[65522]: _type = "Task" [ 1186.247831] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.248071] env[65522]: DEBUG nova.compute.manager [-] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1186.248135] env[65522]: DEBUG nova.network.neutron [-] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1186.248378] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1186.248899] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.249166] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.268909] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114969, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.274631] env[65522]: DEBUG nova.network.neutron [-] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1186.274896] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1186.281691] env[65522]: WARNING neutronclient.v2_0.client [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1186.327756] env[65522]: DEBUG oslo_concurrency.lockutils [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1186.327756] env[65522]: DEBUG oslo_concurrency.lockutils [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1186.327756] env[65522]: DEBUG nova.network.neutron [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1186.421168] env[65522]: DEBUG nova.scheduler.client.report [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1186.652461] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: f6980679-bc2e-442f-b424-8cf80de86c02] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1186.769168] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114969, 'name': ReconfigVM_Task, 'duration_secs': 0.310619} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.769168] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Reconfigured VM instance instance-00000070 to attach disk [datastore1] 3987d71f-47ab-4dd4-8426-95d78503effa/3987d71f-47ab-4dd4-8426-95d78503effa.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1186.770753] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-abefab9a-1c1d-4e73-ac2a-7bdf27a48cad {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.778026] env[65522]: DEBUG nova.network.neutron [-] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1186.780142] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1186.780142] env[65522]: value = "task-5114970" [ 1186.780142] env[65522]: _type = "Task" [ 1186.780142] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.791504] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114970, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.830014] env[65522]: WARNING neutronclient.v2_0.client [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1186.830780] env[65522]: WARNING openstack [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.831283] env[65522]: WARNING openstack [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1186.926607] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.062s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1186.927226] env[65522]: DEBUG nova.compute.manager [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1186.935018] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.204s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1186.937702] env[65522]: INFO nova.compute.claims [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1186.979776] env[65522]: WARNING openstack [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1186.980411] env[65522]: WARNING openstack [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1187.049289] env[65522]: WARNING neutronclient.v2_0.client [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1187.049976] env[65522]: WARNING openstack [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1187.050396] env[65522]: WARNING openstack [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1187.142402] env[65522]: DEBUG nova.network.neutron [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance_info_cache with network_info: [{"id": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "address": "fa:16:3e:5b:f2:d8", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7fc03fa-06", "ovs_interfaceid": "a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1187.155588] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 210184da-48db-4926-ae00-7c0f29622810] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1187.185082] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1187.185359] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1187.185541] env[65522]: INFO nova.compute.manager [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Shelving [ 1187.281485] env[65522]: INFO nova.compute.manager [-] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Took 1.03 seconds to deallocate network for instance. [ 1187.296431] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114970, 'name': Rename_Task, 'duration_secs': 0.162165} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.296706] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1187.296973] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ee76810-c48f-40bf-b910-c4fefc18e19d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.304258] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1187.304258] env[65522]: value = "task-5114972" [ 1187.304258] env[65522]: _type = "Task" [ 1187.304258] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.312910] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114972, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.432389] env[65522]: DEBUG nova.compute.utils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1187.438030] env[65522]: DEBUG nova.compute.manager [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1187.438030] env[65522]: DEBUG nova.network.neutron [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1187.438030] env[65522]: WARNING neutronclient.v2_0.client [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1187.438030] env[65522]: WARNING neutronclient.v2_0.client [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1187.438030] env[65522]: WARNING openstack [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1187.438030] env[65522]: WARNING openstack [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1187.489165] env[65522]: DEBUG nova.policy [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8f8fa586d8c74bb9805a1148fa7ed037', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f93c790ff61543bd8e134bcf9cb20bb2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1187.645255] env[65522]: DEBUG oslo_concurrency.lockutils [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-cb62df50-2f86-469a-b2ee-fd3754d61c83" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1187.645785] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1187.646096] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f19f9a2e-981b-4708-aa23-4989319c4de5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.654196] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1187.654196] env[65522]: value = "task-5114973" [ 1187.654196] env[65522]: _type = "Task" [ 1187.654196] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.659412] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: c071e816-fe74-4aae-9f8e-7a1c69431f9d] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1187.666059] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114973, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.700170] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1187.700426] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1187.700573] env[65522]: INFO nova.compute.manager [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Shelving [ 1187.797639] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1187.816939] env[65522]: DEBUG oslo_vmware.api [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5114972, 'name': PowerOnVM_Task, 'duration_secs': 0.494288} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.817654] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1187.817654] env[65522]: INFO nova.compute.manager [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Took 8.03 seconds to spawn the instance on the hypervisor. [ 1187.817654] env[65522]: DEBUG nova.compute.manager [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1187.818558] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c11f855-f3a8-451c-9c02-48354af6b65b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.845362] env[65522]: DEBUG nova.network.neutron [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Successfully created port: 499e91eb-948a-4b54-b7a6-ce30af8315ea {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1187.945214] env[65522]: DEBUG nova.compute.manager [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1188.089586] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e25df0-6733-4004-aa60-1f19430c7305 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.099120] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbddce39-4675-4c52-91b6-861d7b88641b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.132423] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf95592-9680-4d2a-a258-09a541ad4030 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.140914] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7216b3ad-3bd4-4743-8b6d-0dd9e5043bb4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.155128] env[65522]: DEBUG nova.compute.provider_tree [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1188.165641] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114973, 'name': PowerOffVM_Task, 'duration_secs': 0.22381} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.166144] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1188.166884] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1188.167075] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1188.167484] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1188.167484] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1188.167677] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1188.167677] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1188.167848] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1188.168011] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1188.168184] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1188.168355] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1188.168517] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1188.173659] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: f1ce2d19-347a-4830-8501-8cceb87d4bf8] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1188.176434] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cfbc87d9-0009-48c6-8237-d2b42c7ad9ba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.193857] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1188.193857] env[65522]: value = "task-5114974" [ 1188.193857] env[65522]: _type = "Task" [ 1188.193857] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.198833] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1188.199189] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-86fc8f33-a22a-4146-aa99-90ba3f3074a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.206693] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114974, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.214616] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1188.214616] env[65522]: value = "task-5114975" [ 1188.214616] env[65522]: _type = "Task" [ 1188.214616] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.226623] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114975, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.341375] env[65522]: INFO nova.compute.manager [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Took 18.66 seconds to build instance. [ 1188.661193] env[65522]: DEBUG nova.scheduler.client.report [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1188.688526] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 95f0772d-6d3e-4e37-a0d2-9ab9de63d124] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1188.705310] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114974, 'name': ReconfigVM_Task, 'duration_secs': 0.280185} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.706185] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac3ed8b-b315-4d2a-83ee-1d285bd3c7f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.711018] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1188.711861] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7741897a-7371-4355-a2ec-4568049648d1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.729042] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1188.729042] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1188.729042] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1188.729369] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1188.729369] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1188.729476] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1188.729641] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1188.729806] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1188.729989] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1188.730147] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1188.730321] env[65522]: DEBUG nova.virt.hardware [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1188.735081] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d47e22c-692b-4843-a4e5-c35ecd7524fc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.739391] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1188.739391] env[65522]: value = "task-5114976" [ 1188.739391] env[65522]: _type = "Task" [ 1188.739391] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.748742] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114975, 'name': PowerOffVM_Task, 'duration_secs': 0.201392} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.749064] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1188.749064] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5284e924-ed15-6f3d-d988-0f6cef68d0df" [ 1188.749064] env[65522]: _type = "Task" [ 1188.749064] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.749693] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1188.750597] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe602f3-0636-4d88-9f12-4ddba58faddf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.760268] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114976, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.766488] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5284e924-ed15-6f3d-d988-0f6cef68d0df, 'name': SearchDatastore_Task, 'duration_secs': 0.008678} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.786087] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Reconfiguring VM instance instance-00000064 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1188.786999] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd598753-e5be-4ca9-988a-7fe86f893ae0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.801302] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca6876b-adc4-450f-816f-76a02dc4bbac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.813535] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1188.813535] env[65522]: value = "task-5114977" [ 1188.813535] env[65522]: _type = "Task" [ 1188.813535] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.823914] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114977, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.843800] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e96cba8-8107-4db5-b9ba-1f9ba2c24ed6 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.175s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1188.964910] env[65522]: DEBUG nova.compute.manager [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1188.997465] env[65522]: DEBUG nova.virt.hardware [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1188.997726] env[65522]: DEBUG nova.virt.hardware [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1188.997880] env[65522]: DEBUG nova.virt.hardware [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1188.998073] env[65522]: DEBUG nova.virt.hardware [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1188.998219] env[65522]: DEBUG nova.virt.hardware [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1188.998365] env[65522]: DEBUG nova.virt.hardware [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1188.998577] env[65522]: DEBUG nova.virt.hardware [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1188.998800] env[65522]: DEBUG nova.virt.hardware [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1188.998975] env[65522]: DEBUG nova.virt.hardware [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1188.999153] env[65522]: DEBUG nova.virt.hardware [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1188.999342] env[65522]: DEBUG nova.virt.hardware [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1189.000295] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e93d6cd-5cdb-4cf7-8909-56dc0aab0332 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.009576] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60694a7c-9e0f-495b-b0c5-7eb72c725dc9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.166788] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.231s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.167377] env[65522]: DEBUG nova.compute.manager [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1189.170070] env[65522]: DEBUG oslo_concurrency.lockutils [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 6.696s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1189.170269] env[65522]: DEBUG nova.objects.instance [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1189.193039] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: b412d3e5-0536-413c-9519-5f85c7647cdf] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1189.253383] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114976, 'name': PowerOffVM_Task, 'duration_secs': 0.226261} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.254229] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1189.255254] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef36059-ac03-46e8-8e1b-4e2555cc4e6a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.282154] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc5f8d67-3c40-48c0-afac-5276a14214eb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.315183] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1189.315739] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c3a3d91e-cccb-4dd4-9732-0041bd8ffc52 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.328889] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114977, 'name': ReconfigVM_Task, 'duration_secs': 0.206999} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.330545] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Reconfigured VM instance instance-00000064 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1189.330947] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1189.330947] env[65522]: value = "task-5114979" [ 1189.330947] env[65522]: _type = "Task" [ 1189.330947] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.332022] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88b966f-a6f9-4b06-a5db-f142cb3c9265 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.343764] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114979, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.362323] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] cb62df50-2f86-469a-b2ee-fd3754d61c83/cb62df50-2f86-469a-b2ee-fd3754d61c83.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1189.362675] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95852786-b1fe-4840-9528-93ac8f1d25dc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.380518] env[65522]: DEBUG nova.compute.manager [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Received event network-changed-c8c5ef6a-c86a-4a39-a2cc-918754e4e78a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1189.380800] env[65522]: DEBUG nova.compute.manager [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Refreshing instance network info cache due to event network-changed-c8c5ef6a-c86a-4a39-a2cc-918754e4e78a. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1189.380904] env[65522]: DEBUG oslo_concurrency.lockutils [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] Acquiring lock "refresh_cache-3987d71f-47ab-4dd4-8426-95d78503effa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.381064] env[65522]: DEBUG oslo_concurrency.lockutils [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] Acquired lock "refresh_cache-3987d71f-47ab-4dd4-8426-95d78503effa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1189.381249] env[65522]: DEBUG nova.network.neutron [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Refreshing network info cache for port c8c5ef6a-c86a-4a39-a2cc-918754e4e78a {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1189.385694] env[65522]: DEBUG nova.compute.manager [req-404cc580-e0c5-4fc5-8430-6378772f07e4 req-1d38fbd7-c80b-4bc1-8a33-1655265dd78f service nova] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Received event network-vif-plugged-499e91eb-948a-4b54-b7a6-ce30af8315ea {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1189.385907] env[65522]: DEBUG oslo_concurrency.lockutils [req-404cc580-e0c5-4fc5-8430-6378772f07e4 req-1d38fbd7-c80b-4bc1-8a33-1655265dd78f service nova] Acquiring lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1189.386123] env[65522]: DEBUG oslo_concurrency.lockutils [req-404cc580-e0c5-4fc5-8430-6378772f07e4 req-1d38fbd7-c80b-4bc1-8a33-1655265dd78f service nova] Lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1189.386287] env[65522]: DEBUG oslo_concurrency.lockutils [req-404cc580-e0c5-4fc5-8430-6378772f07e4 req-1d38fbd7-c80b-4bc1-8a33-1655265dd78f service nova] Lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.386448] env[65522]: DEBUG nova.compute.manager [req-404cc580-e0c5-4fc5-8430-6378772f07e4 req-1d38fbd7-c80b-4bc1-8a33-1655265dd78f service nova] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] No waiting events found dispatching network-vif-plugged-499e91eb-948a-4b54-b7a6-ce30af8315ea {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1189.386606] env[65522]: WARNING nova.compute.manager [req-404cc580-e0c5-4fc5-8430-6378772f07e4 req-1d38fbd7-c80b-4bc1-8a33-1655265dd78f service nova] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Received unexpected event network-vif-plugged-499e91eb-948a-4b54-b7a6-ce30af8315ea for instance with vm_state building and task_state spawning. [ 1189.391678] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1189.391678] env[65522]: value = "task-5114980" [ 1189.391678] env[65522]: _type = "Task" [ 1189.391678] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.404649] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114980, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.438192] env[65522]: DEBUG nova.network.neutron [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Successfully updated port: 499e91eb-948a-4b54-b7a6-ce30af8315ea {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1189.675157] env[65522]: DEBUG nova.compute.utils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1189.679272] env[65522]: DEBUG nova.compute.manager [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1189.679469] env[65522]: DEBUG nova.network.neutron [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1189.679823] env[65522]: WARNING neutronclient.v2_0.client [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1189.680137] env[65522]: WARNING neutronclient.v2_0.client [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1189.680829] env[65522]: WARNING openstack [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1189.681066] env[65522]: WARNING openstack [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1189.689742] env[65522]: DEBUG oslo_concurrency.lockutils [None req-29d54eae-db3f-4fbf-9987-03b07a5083b7 tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.520s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.691566] env[65522]: DEBUG nova.compute.manager [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1189.694343] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.731s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1189.696194] env[65522]: INFO nova.compute.claims [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1189.700713] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 7a358133-88a5-4dd7-ab69-45b8172870a5] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1189.749674] env[65522]: DEBUG nova.policy [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2f843db470614ed9b09fb5b6463827e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '15229af4da7a4ff9963369df50fb9d5b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1189.795075] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1189.795517] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1389dab1-8dd6-4b1a-9dbd-b322e806e7df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.804328] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1189.804328] env[65522]: value = "task-5114981" [ 1189.804328] env[65522]: _type = "Task" [ 1189.804328] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.813515] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114981, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.845596] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114979, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.884567] env[65522]: WARNING neutronclient.v2_0.client [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1189.885286] env[65522]: WARNING openstack [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1189.885671] env[65522]: WARNING openstack [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1189.902439] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114980, 'name': ReconfigVM_Task, 'duration_secs': 0.320634} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.902754] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Reconfigured VM instance instance-00000064 to attach disk [datastore2] cb62df50-2f86-469a-b2ee-fd3754d61c83/cb62df50-2f86-469a-b2ee-fd3754d61c83.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1189.903691] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed3c6bc-9a41-4885-940e-d3db690f1b7e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.923783] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58653433-b0ca-461f-91e5-e5479c3805ae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.945260] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.945449] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1189.945634] env[65522]: DEBUG nova.network.neutron [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1189.948684] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c1448e-60b3-4e85-a486-73fb47ff973f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.975674] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f9bcbe-a919-481b-ae9f-a6c51028ebd1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.986443] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1189.986755] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0ded4ac-04c3-4493-8392-6e64bef64db9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.994142] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1189.994142] env[65522]: value = "task-5114982" [ 1189.994142] env[65522]: _type = "Task" [ 1189.994142] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.003682] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114982, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.021362] env[65522]: WARNING openstack [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1190.023620] env[65522]: WARNING openstack [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1190.069697] env[65522]: DEBUG nova.network.neutron [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Successfully created port: 54d3b471-cec2-4c58-a51d-7bd3ed42e50e {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1190.087527] env[65522]: WARNING neutronclient.v2_0.client [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1190.088433] env[65522]: WARNING openstack [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1190.088771] env[65522]: WARNING openstack [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1190.175158] env[65522]: DEBUG nova.network.neutron [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Updated VIF entry in instance network info cache for port c8c5ef6a-c86a-4a39-a2cc-918754e4e78a. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1190.175538] env[65522]: DEBUG nova.network.neutron [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Updating instance_info_cache with network_info: [{"id": "c8c5ef6a-c86a-4a39-a2cc-918754e4e78a", "address": "fa:16:3e:ee:34:76", "network": {"id": "f49e9b61-6927-4a5b-a12d-f54f408d42b4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1397158171-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a80f979f3dc0477e9462b47f7aa87f14", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc8c5ef6a-c8", "ovs_interfaceid": "c8c5ef6a-c86a-4a39-a2cc-918754e4e78a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1190.210220] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 99353848-2f0f-4388-9fcd-91e799342386] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1190.317328] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114981, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.351381] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114979, 'name': CreateSnapshot_Task, 'duration_secs': 0.66409} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.351732] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1190.352947] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bcb6aac-8996-4334-b697-b9acbf056249 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.452479] env[65522]: WARNING openstack [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1190.452886] env[65522]: WARNING openstack [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1190.507868] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114982, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.584250] env[65522]: DEBUG nova.network.neutron [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1190.625162] env[65522]: WARNING openstack [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1190.625634] env[65522]: WARNING openstack [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1190.679599] env[65522]: DEBUG oslo_concurrency.lockutils [req-5fcf8e9f-e74b-4cc9-8356-3e956f010cb4 req-dc8a3f7b-7088-4f53-bc44-cf7787f27433 service nova] Releasing lock "refresh_cache-3987d71f-47ab-4dd4-8426-95d78503effa" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1190.706378] env[65522]: DEBUG nova.compute.manager [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1190.720228] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 43691901-7ac0-4523-a19c-eedf5ae4c040] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1190.724221] env[65522]: WARNING neutronclient.v2_0.client [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1190.724451] env[65522]: WARNING openstack [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1190.724820] env[65522]: WARNING openstack [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1190.742268] env[65522]: DEBUG nova.virt.hardware [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1190.742585] env[65522]: DEBUG nova.virt.hardware [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1190.742771] env[65522]: DEBUG nova.virt.hardware [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1190.742967] env[65522]: DEBUG nova.virt.hardware [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1190.743131] env[65522]: DEBUG nova.virt.hardware [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1190.743338] env[65522]: DEBUG nova.virt.hardware [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1190.743547] env[65522]: DEBUG nova.virt.hardware [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1190.743778] env[65522]: DEBUG nova.virt.hardware [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1190.744090] env[65522]: DEBUG nova.virt.hardware [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1190.744330] env[65522]: DEBUG nova.virt.hardware [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1190.744622] env[65522]: DEBUG nova.virt.hardware [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1190.745701] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61759e8-9383-48ac-bd39-fb6e604baf46 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.759987] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf0af2a-d45c-4d5a-baae-02cc176b151a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.818093] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114981, 'name': CreateSnapshot_Task, 'duration_secs': 0.779387} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.818458] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1190.819249] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d683bc-2747-4c94-94ec-6de63028c33c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.823812] env[65522]: DEBUG nova.network.neutron [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating instance_info_cache with network_info: [{"id": "499e91eb-948a-4b54-b7a6-ce30af8315ea", "address": "fa:16:3e:09:07:e8", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap499e91eb-94", "ovs_interfaceid": "499e91eb-948a-4b54-b7a6-ce30af8315ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1190.878561] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1190.880609] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8c68631b-10eb-4fc3-9c86-a8b3cdc6c2e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.884471] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-896c6537-af77-4321-84f2-4964e0e0db44 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.894392] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cddd2e4c-dbed-4c0e-ae24-a55467c86296 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.897608] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1190.897608] env[65522]: value = "task-5114983" [ 1190.897608] env[65522]: _type = "Task" [ 1190.897608] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.930689] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af1d7f6-d625-4246-937b-ea721b87f71f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.933593] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114983, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.940457] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10bf8ddc-5dc3-4459-84ca-8d09783a53fb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.956464] env[65522]: DEBUG nova.compute.provider_tree [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1191.007195] env[65522]: DEBUG oslo_vmware.api [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114982, 'name': PowerOnVM_Task, 'duration_secs': 0.544219} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.007544] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1191.236280] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 5cb16f07-3bad-4e69-85e5-5fa4fc0b66a7] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1191.326939] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1191.327447] env[65522]: DEBUG nova.compute.manager [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Instance network_info: |[{"id": "499e91eb-948a-4b54-b7a6-ce30af8315ea", "address": "fa:16:3e:09:07:e8", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap499e91eb-94", "ovs_interfaceid": "499e91eb-948a-4b54-b7a6-ce30af8315ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1191.327937] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:07:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '499e91eb-948a-4b54-b7a6-ce30af8315ea', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1191.337247] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1191.345885] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1191.346252] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1191.346503] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-be7a6ccf-f3b7-4b59-b92e-ecc092151dfa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.349699] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8036688a-b2e4-4067-b7fc-dd5396f58ac7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.374831] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1191.374831] env[65522]: value = "task-5114985" [ 1191.374831] env[65522]: _type = "Task" [ 1191.374831] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.376343] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1191.376343] env[65522]: value = "task-5114986" [ 1191.376343] env[65522]: _type = "Task" [ 1191.376343] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.393072] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114985, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.397166] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114986, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.410539] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114983, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.460739] env[65522]: DEBUG nova.scheduler.client.report [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1191.625390] env[65522]: DEBUG nova.compute.manager [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Received event network-changed-499e91eb-948a-4b54-b7a6-ce30af8315ea {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1191.625667] env[65522]: DEBUG nova.compute.manager [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Refreshing instance network info cache due to event network-changed-499e91eb-948a-4b54-b7a6-ce30af8315ea. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1191.625933] env[65522]: DEBUG oslo_concurrency.lockutils [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] Acquiring lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.626126] env[65522]: DEBUG oslo_concurrency.lockutils [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] Acquired lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1191.626349] env[65522]: DEBUG nova.network.neutron [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Refreshing network info cache for port 499e91eb-948a-4b54-b7a6-ce30af8315ea {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1191.712706] env[65522]: DEBUG nova.network.neutron [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Successfully updated port: 54d3b471-cec2-4c58-a51d-7bd3ed42e50e {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1191.740316] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: e7baf2a0-21dd-4610-8230-81b99b64856b] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1191.892941] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114985, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.897357] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114986, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.909806] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114983, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.968329] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.274s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1191.968996] env[65522]: DEBUG nova.compute.manager [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1191.972879] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.175s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1191.973154] env[65522]: DEBUG nova.objects.instance [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Lazy-loading 'resources' on Instance uuid 6d856111-81bd-4e2f-b7fc-e169c3e974a6 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1192.021319] env[65522]: INFO nova.compute.manager [None req-87436b92-2f3d-4222-911c-a3fce1c8078a tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance to original state: 'active' [ 1192.129482] env[65522]: WARNING neutronclient.v2_0.client [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1192.130464] env[65522]: WARNING openstack [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1192.130653] env[65522]: WARNING openstack [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1192.219255] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "refresh_cache-874b6a07-ebba-4f75-937e-2281af1598b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.219343] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquired lock "refresh_cache-874b6a07-ebba-4f75-937e-2281af1598b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1192.219524] env[65522]: DEBUG nova.network.neutron [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1192.232049] env[65522]: WARNING openstack [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1192.232439] env[65522]: WARNING openstack [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1192.243914] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 11ef190f-8ade-4705-b3b0-a8ff8b97bcd3] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1192.301653] env[65522]: WARNING neutronclient.v2_0.client [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1192.302376] env[65522]: WARNING openstack [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1192.302736] env[65522]: WARNING openstack [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1192.391720] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114985, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.396447] env[65522]: DEBUG nova.network.neutron [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updated VIF entry in instance network info cache for port 499e91eb-948a-4b54-b7a6-ce30af8315ea. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1192.396809] env[65522]: DEBUG nova.network.neutron [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating instance_info_cache with network_info: [{"id": "499e91eb-948a-4b54-b7a6-ce30af8315ea", "address": "fa:16:3e:09:07:e8", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap499e91eb-94", "ovs_interfaceid": "499e91eb-948a-4b54-b7a6-ce30af8315ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1192.401366] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114986, 'name': CreateVM_Task, 'duration_secs': 0.568356} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.404953] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1192.405529] env[65522]: WARNING neutronclient.v2_0.client [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1192.405857] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.406018] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1192.406330] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1192.407570] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e5223b8-1f90-41dc-83d9-2cc63453ce72 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.414196] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5114983, 'name': CloneVM_Task, 'duration_secs': 1.278841} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.415489] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Created linked-clone VM from snapshot [ 1192.415852] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1192.415852] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5201d5b3-129e-9087-3663-b0f3283d7741" [ 1192.415852] env[65522]: _type = "Task" [ 1192.415852] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.416584] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fa04749-b320-42ab-8d16-705772ef4935 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.428498] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Uploading image 7832ad6e-429a-4894-9720-58fd3520c4e2 {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1192.434568] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5201d5b3-129e-9087-3663-b0f3283d7741, 'name': SearchDatastore_Task, 'duration_secs': 0.011738} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.434884] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1192.435371] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1192.435371] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.435506] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1192.435665] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1192.436289] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e84ab4b9-dce7-4293-84de-185e95f49e5b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.446905] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1192.447189] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1192.447962] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-456e97f3-6c77-4042-a869-89307027e094 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.456845] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1192.456845] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521b0399-3b56-8ad8-80fb-7411fcbd008a" [ 1192.456845] env[65522]: _type = "Task" [ 1192.456845] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.462072] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1192.462072] env[65522]: value = "vm-994960" [ 1192.462072] env[65522]: _type = "VirtualMachine" [ 1192.462072] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1192.462341] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-27a5513f-63df-46df-bd43-0f89f422eb34 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.470337] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521b0399-3b56-8ad8-80fb-7411fcbd008a, 'name': SearchDatastore_Task, 'duration_secs': 0.011255} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.472446] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lease: (returnval){ [ 1192.472446] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52896382-a915-25ba-84d2-5e28e03a8fe1" [ 1192.472446] env[65522]: _type = "HttpNfcLease" [ 1192.472446] env[65522]: } obtained for exporting VM: (result){ [ 1192.472446] env[65522]: value = "vm-994960" [ 1192.472446] env[65522]: _type = "VirtualMachine" [ 1192.472446] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1192.472728] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the lease: (returnval){ [ 1192.472728] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52896382-a915-25ba-84d2-5e28e03a8fe1" [ 1192.472728] env[65522]: _type = "HttpNfcLease" [ 1192.472728] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1192.472917] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3875c9e7-97ed-49a5-b983-4c329ff1a1da {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.482624] env[65522]: DEBUG nova.compute.utils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1192.486449] env[65522]: DEBUG nova.compute.manager [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1192.486696] env[65522]: DEBUG nova.network.neutron [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1192.487151] env[65522]: WARNING neutronclient.v2_0.client [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1192.487388] env[65522]: WARNING neutronclient.v2_0.client [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1192.488062] env[65522]: WARNING openstack [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1192.488425] env[65522]: WARNING openstack [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1192.496733] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1192.496733] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52907bf8-f62f-c7c1-7465-5e3b3fc65721" [ 1192.496733] env[65522]: _type = "Task" [ 1192.496733] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.497923] env[65522]: DEBUG nova.compute.manager [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1192.507710] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1192.507710] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52896382-a915-25ba-84d2-5e28e03a8fe1" [ 1192.507710] env[65522]: _type = "HttpNfcLease" [ 1192.507710] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1192.511198] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1192.511198] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52896382-a915-25ba-84d2-5e28e03a8fe1" [ 1192.511198] env[65522]: _type = "HttpNfcLease" [ 1192.511198] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1192.512744] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ae3268-b4fe-42d3-b514-00dbed3281e3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.521133] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52907bf8-f62f-c7c1-7465-5e3b3fc65721, 'name': SearchDatastore_Task, 'duration_secs': 0.010827} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.522057] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1192.522183] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] ae3455d8-2f38-42ad-b16d-d98aef92b2a8/ae3455d8-2f38-42ad-b16d-d98aef92b2a8.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1192.522526] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8767968a-7f24-463d-add3-6bf8586951f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.537012] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d9517-44f9-afc3-b6a1-d09bd16ad6fa/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1192.537012] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d9517-44f9-afc3-b6a1-d09bd16ad6fa/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1192.605079] env[65522]: DEBUG nova.policy [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b51f1b47ad744dc2b2b0598cee11bc6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d1239b79ae94cceb89ae7a8bd57da08', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1192.611552] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1192.611552] env[65522]: value = "task-5114988" [ 1192.611552] env[65522]: _type = "Task" [ 1192.611552] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.626773] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114988, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.664319] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-02a48352-5de4-4d67-9a95-34cf5548b214 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.722824] env[65522]: WARNING openstack [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1192.723359] env[65522]: WARNING openstack [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1192.747570] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 8859b051-8f75-4aad-b789-42662019d4c5] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1192.753434] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce675939-152c-47f6-9b0b-fd17906f1300 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.769607] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb04502-f7ce-4b04-9af1-8e5e8b9d40a6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.808163] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ed873a-8a93-475f-88ce-179469d7769d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.821176] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca010b3-7d49-4cd9-945c-95ece5120b27 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.826546] env[65522]: DEBUG nova.network.neutron [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1192.843137] env[65522]: DEBUG nova.compute.provider_tree [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1192.891447] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114985, 'name': CloneVM_Task} progress is 95%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.895463] env[65522]: WARNING openstack [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1192.895918] env[65522]: WARNING openstack [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1192.903276] env[65522]: DEBUG oslo_concurrency.lockutils [req-0a04ee88-0729-4f15-bc02-760fb8d7e34c req-53ab4862-0104-4a3f-a0e3-1271ffaaeb85 service nova] Releasing lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1193.076545] env[65522]: DEBUG nova.network.neutron [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Successfully created port: 4a2d504c-073b-4b0a-b208-065c99308adc {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1193.091070] env[65522]: WARNING neutronclient.v2_0.client [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1193.091070] env[65522]: WARNING openstack [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1193.091452] env[65522]: WARNING openstack [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1193.127801] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114988, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.259543] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 00fcbbcb-ef87-4318-8c6e-ce62feb9fd83] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1193.287623] env[65522]: DEBUG nova.network.neutron [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Updating instance_info_cache with network_info: [{"id": "54d3b471-cec2-4c58-a51d-7bd3ed42e50e", "address": "fa:16:3e:08:2f:12", "network": {"id": "26dda233-0dd0-40ae-b56f-c84b1b91e92a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1197499239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "15229af4da7a4ff9963369df50fb9d5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54d3b471-ce", "ovs_interfaceid": "54d3b471-cec2-4c58-a51d-7bd3ed42e50e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1193.348250] env[65522]: DEBUG nova.scheduler.client.report [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1193.393042] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5114985, 'name': CloneVM_Task, 'duration_secs': 1.938615} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.393598] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Created linked-clone VM from snapshot [ 1193.394798] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96690d91-8d0d-4f71-aba1-66d92bd20109 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.405186] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Uploading image e4e8fd5f-0bc1-437d-a8db-457c5ea05344 {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1193.433838] env[65522]: DEBUG oslo_vmware.rw_handles [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1193.433838] env[65522]: value = "vm-994963" [ 1193.433838] env[65522]: _type = "VirtualMachine" [ 1193.433838] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1193.435494] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9d4123c5-16a9-42cb-97de-377163e56e63 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.445347] env[65522]: DEBUG oslo_vmware.rw_handles [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lease: (returnval){ [ 1193.445347] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d072cd-b722-6ca3-0227-d64869070ee9" [ 1193.445347] env[65522]: _type = "HttpNfcLease" [ 1193.445347] env[65522]: } obtained for exporting VM: (result){ [ 1193.445347] env[65522]: value = "vm-994963" [ 1193.445347] env[65522]: _type = "VirtualMachine" [ 1193.445347] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1193.446360] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the lease: (returnval){ [ 1193.446360] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d072cd-b722-6ca3-0227-d64869070ee9" [ 1193.446360] env[65522]: _type = "HttpNfcLease" [ 1193.446360] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1193.455284] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1193.455284] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d072cd-b722-6ca3-0227-d64869070ee9" [ 1193.455284] env[65522]: _type = "HttpNfcLease" [ 1193.455284] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1193.515021] env[65522]: DEBUG nova.compute.manager [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1193.543878] env[65522]: DEBUG nova.virt.hardware [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1193.544438] env[65522]: DEBUG nova.virt.hardware [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1193.544733] env[65522]: DEBUG nova.virt.hardware [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1193.545015] env[65522]: DEBUG nova.virt.hardware [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1193.545220] env[65522]: DEBUG nova.virt.hardware [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1193.545731] env[65522]: DEBUG nova.virt.hardware [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1193.545731] env[65522]: DEBUG nova.virt.hardware [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1193.545894] env[65522]: DEBUG nova.virt.hardware [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1193.546083] env[65522]: DEBUG nova.virt.hardware [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1193.546365] env[65522]: DEBUG nova.virt.hardware [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1193.546624] env[65522]: DEBUG nova.virt.hardware [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1193.547942] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773a816d-4d9d-470b-8587-f2dd65e9f0d0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.558203] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee164172-e066-402f-8e46-48aa386b7cb5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.625984] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114988, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.667738] env[65522]: DEBUG nova.compute.manager [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Received event network-vif-plugged-54d3b471-cec2-4c58-a51d-7bd3ed42e50e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1193.667814] env[65522]: DEBUG oslo_concurrency.lockutils [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] Acquiring lock "874b6a07-ebba-4f75-937e-2281af1598b8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1193.668098] env[65522]: DEBUG oslo_concurrency.lockutils [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] Lock "874b6a07-ebba-4f75-937e-2281af1598b8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1193.668354] env[65522]: DEBUG oslo_concurrency.lockutils [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] Lock "874b6a07-ebba-4f75-937e-2281af1598b8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1193.668535] env[65522]: DEBUG nova.compute.manager [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] No waiting events found dispatching network-vif-plugged-54d3b471-cec2-4c58-a51d-7bd3ed42e50e {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1193.668708] env[65522]: WARNING nova.compute.manager [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Received unexpected event network-vif-plugged-54d3b471-cec2-4c58-a51d-7bd3ed42e50e for instance with vm_state building and task_state spawning. [ 1193.669143] env[65522]: DEBUG nova.compute.manager [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Received event network-changed-54d3b471-cec2-4c58-a51d-7bd3ed42e50e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1193.670077] env[65522]: DEBUG nova.compute.manager [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Refreshing instance network info cache due to event network-changed-54d3b471-cec2-4c58-a51d-7bd3ed42e50e. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1193.670314] env[65522]: DEBUG oslo_concurrency.lockutils [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] Acquiring lock "refresh_cache-874b6a07-ebba-4f75-937e-2281af1598b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.768850] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 7701b92d-805e-4837-b900-326910b5eef5] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1193.791514] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Releasing lock "refresh_cache-874b6a07-ebba-4f75-937e-2281af1598b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1193.791910] env[65522]: DEBUG nova.compute.manager [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Instance network_info: |[{"id": "54d3b471-cec2-4c58-a51d-7bd3ed42e50e", "address": "fa:16:3e:08:2f:12", "network": {"id": "26dda233-0dd0-40ae-b56f-c84b1b91e92a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1197499239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "15229af4da7a4ff9963369df50fb9d5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54d3b471-ce", "ovs_interfaceid": "54d3b471-cec2-4c58-a51d-7bd3ed42e50e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1193.792272] env[65522]: DEBUG oslo_concurrency.lockutils [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] Acquired lock "refresh_cache-874b6a07-ebba-4f75-937e-2281af1598b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1193.792525] env[65522]: DEBUG nova.network.neutron [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Refreshing network info cache for port 54d3b471-cec2-4c58-a51d-7bd3ed42e50e {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1193.794029] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:2f:12', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa8c2f93-f287-41b3-adb6-4942a7ea2a0b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '54d3b471-cec2-4c58-a51d-7bd3ed42e50e', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1193.802461] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Creating folder: Project (15229af4da7a4ff9963369df50fb9d5b). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1193.804263] env[65522]: WARNING neutronclient.v2_0.client [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1193.804913] env[65522]: WARNING openstack [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1193.805610] env[65522]: WARNING openstack [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1193.814148] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c71a82c5-777b-42eb-88bc-d7db05ec8b22 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.827842] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Created folder: Project (15229af4da7a4ff9963369df50fb9d5b) in parent group-v994660. [ 1193.828224] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Creating folder: Instances. Parent ref: group-v994964. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1193.828298] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-28048984-ba64-4080-b898-54503713abfe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.839532] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Created folder: Instances in parent group-v994964. [ 1193.839907] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1193.840155] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1193.840502] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-36ba00b1-ea85-4203-a4c0-b5c7c1338a04 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.856936] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.884s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1193.866362] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1193.866362] env[65522]: value = "task-5114993" [ 1193.866362] env[65522]: _type = "Task" [ 1193.866362] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.880504] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114993, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.884836] env[65522]: INFO nova.scheduler.client.report [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Deleted allocations for instance 6d856111-81bd-4e2f-b7fc-e169c3e974a6 [ 1193.959411] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1193.959411] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d072cd-b722-6ca3-0227-d64869070ee9" [ 1193.959411] env[65522]: _type = "HttpNfcLease" [ 1193.959411] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1193.959925] env[65522]: DEBUG oslo_vmware.rw_handles [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1193.959925] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d072cd-b722-6ca3-0227-d64869070ee9" [ 1193.959925] env[65522]: _type = "HttpNfcLease" [ 1193.959925] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1193.961313] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e64609a-496c-4fcc-a42f-8735e2545ca9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.965768] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "cb62df50-2f86-469a-b2ee-fd3754d61c83" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1193.966177] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "cb62df50-2f86-469a-b2ee-fd3754d61c83" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1193.966474] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "cb62df50-2f86-469a-b2ee-fd3754d61c83-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1193.966718] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "cb62df50-2f86-469a-b2ee-fd3754d61c83-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1193.966908] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "cb62df50-2f86-469a-b2ee-fd3754d61c83-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1193.969383] env[65522]: INFO nova.compute.manager [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Terminating instance [ 1193.982435] env[65522]: DEBUG oslo_vmware.rw_handles [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522d7da7-f0e0-d089-e742-411604bc1ffb/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1193.982672] env[65522]: DEBUG oslo_vmware.rw_handles [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522d7da7-f0e0-d089-e742-411604bc1ffb/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1194.061316] env[65522]: WARNING openstack [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1194.061776] env[65522]: WARNING openstack [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1194.126246] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114988, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.557381} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.126866] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] ae3455d8-2f38-42ad-b16d-d98aef92b2a8/ae3455d8-2f38-42ad-b16d-d98aef92b2a8.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1194.127109] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1194.127552] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-55f137a0-2131-4c03-9223-54eed378cb35 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.135704] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1194.135704] env[65522]: value = "task-5114994" [ 1194.135704] env[65522]: _type = "Task" [ 1194.135704] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.147803] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e08eac02-b4bc-44a8-bc69-28da48f90d0b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.159150] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114994, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.202624] env[65522]: WARNING neutronclient.v2_0.client [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1194.203603] env[65522]: WARNING openstack [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1194.204252] env[65522]: WARNING openstack [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1194.276993] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 95a10a6f-8c6c-435d-a360-dff10eee0855] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1194.342307] env[65522]: DEBUG nova.network.neutron [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Updated VIF entry in instance network info cache for port 54d3b471-cec2-4c58-a51d-7bd3ed42e50e. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1194.343458] env[65522]: DEBUG nova.network.neutron [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Updating instance_info_cache with network_info: [{"id": "54d3b471-cec2-4c58-a51d-7bd3ed42e50e", "address": "fa:16:3e:08:2f:12", "network": {"id": "26dda233-0dd0-40ae-b56f-c84b1b91e92a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1197499239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "15229af4da7a4ff9963369df50fb9d5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54d3b471-ce", "ovs_interfaceid": "54d3b471-cec2-4c58-a51d-7bd3ed42e50e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1194.381838] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114993, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.394151] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4b5991f7-485b-4482-b8e2-8a4c6578657e tempest-ServersListShow2100Test-36090429 tempest-ServersListShow2100Test-36090429-project-member] Lock "6d856111-81bd-4e2f-b7fc-e169c3e974a6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.431s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1194.486886] env[65522]: DEBUG nova.compute.manager [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1194.487177] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1194.488259] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69012566-a439-42f0-9ca2-0eecb6c27fa2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.498762] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1194.499188] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64c71014-08c4-4caa-bb49-a004325d5fa7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.508295] env[65522]: DEBUG oslo_vmware.api [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1194.508295] env[65522]: value = "task-5114995" [ 1194.508295] env[65522]: _type = "Task" [ 1194.508295] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.521903] env[65522]: DEBUG oslo_vmware.api [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114995, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.650183] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114994, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.137414} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.650709] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1194.651786] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25606f3a-3cc6-4604-979a-3c1035248731 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.694514] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] ae3455d8-2f38-42ad-b16d-d98aef92b2a8/ae3455d8-2f38-42ad-b16d-d98aef92b2a8.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1194.695340] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ccf0e0e-009c-4ed5-9c4a-3cdc7afa795e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.732587] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1194.732587] env[65522]: value = "task-5114996" [ 1194.732587] env[65522]: _type = "Task" [ 1194.732587] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.746608] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114996, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.776918] env[65522]: DEBUG nova.network.neutron [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Successfully updated port: 4a2d504c-073b-4b0a-b208-065c99308adc {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1194.785546] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 82a7ad85-a061-4eec-bd3f-fc977532dfbf] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1194.848596] env[65522]: DEBUG oslo_concurrency.lockutils [req-43444047-ac01-4ecf-9578-bb0830eeadd4 req-5329bdb9-a560-4787-b739-f1edd9088575 service nova] Releasing lock "refresh_cache-874b6a07-ebba-4f75-937e-2281af1598b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1194.879208] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114993, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.026696] env[65522]: DEBUG oslo_vmware.api [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114995, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.247019] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114996, 'name': ReconfigVM_Task, 'duration_secs': 0.461538} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.247544] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Reconfigured VM instance instance-00000071 to attach disk [datastore2] ae3455d8-2f38-42ad-b16d-d98aef92b2a8/ae3455d8-2f38-42ad-b16d-d98aef92b2a8.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1195.248493] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1271cf31-bfc0-4e3a-84bc-9ee6d03f2b2f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.258068] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1195.258068] env[65522]: value = "task-5114998" [ 1195.258068] env[65522]: _type = "Task" [ 1195.258068] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.270199] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114998, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.288392] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.288392] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1195.288684] env[65522]: DEBUG nova.network.neutron [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1195.291077] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 1b2779f5-c6e8-4226-a819-0560d63bd7df] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1195.380682] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114993, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.525686] env[65522]: DEBUG oslo_vmware.api [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114995, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.707762] env[65522]: DEBUG nova.compute.manager [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Received event network-vif-plugged-4a2d504c-073b-4b0a-b208-065c99308adc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1195.707991] env[65522]: DEBUG oslo_concurrency.lockutils [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] Acquiring lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1195.709124] env[65522]: DEBUG oslo_concurrency.lockutils [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.709334] env[65522]: DEBUG oslo_concurrency.lockutils [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1195.709996] env[65522]: DEBUG nova.compute.manager [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] No waiting events found dispatching network-vif-plugged-4a2d504c-073b-4b0a-b208-065c99308adc {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1195.709996] env[65522]: WARNING nova.compute.manager [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Received unexpected event network-vif-plugged-4a2d504c-073b-4b0a-b208-065c99308adc for instance with vm_state building and task_state spawning. [ 1195.709996] env[65522]: DEBUG nova.compute.manager [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Received event network-changed-4a2d504c-073b-4b0a-b208-065c99308adc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1195.710486] env[65522]: DEBUG nova.compute.manager [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Refreshing instance network info cache due to event network-changed-4a2d504c-073b-4b0a-b208-065c99308adc. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1195.710486] env[65522]: DEBUG oslo_concurrency.lockutils [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] Acquiring lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.768926] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114998, 'name': Rename_Task, 'duration_secs': 0.243547} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.769341] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1195.769667] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d3644a2-aa84-4da8-b5f6-47b365bd9310 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.777196] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1195.777196] env[65522]: value = "task-5114999" [ 1195.777196] env[65522]: _type = "Task" [ 1195.777196] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.787386] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114999, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.794722] env[65522]: WARNING openstack [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1195.795306] env[65522]: WARNING openstack [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1195.803961] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 79dd37b5-6b30-48ab-9f00-78214cbd132d] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1195.842692] env[65522]: DEBUG nova.network.neutron [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1195.863561] env[65522]: WARNING openstack [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1195.864280] env[65522]: WARNING openstack [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1195.881839] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5114993, 'name': CreateVM_Task, 'duration_secs': 1.597268} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.882013] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1195.882768] env[65522]: WARNING neutronclient.v2_0.client [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1195.883042] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.883371] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1195.883691] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1195.884096] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b7a58f5-2883-421f-9327-38ae5f67e3ce {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.890323] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1195.890323] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52325c92-b6b6-2a50-348e-901cf948bfd0" [ 1195.890323] env[65522]: _type = "Task" [ 1195.890323] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.900621] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52325c92-b6b6-2a50-348e-901cf948bfd0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.941872] env[65522]: WARNING neutronclient.v2_0.client [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1195.942648] env[65522]: WARNING openstack [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1195.942930] env[65522]: WARNING openstack [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.024656] env[65522]: DEBUG oslo_vmware.api [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5114995, 'name': PowerOffVM_Task, 'duration_secs': 1.105417} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.025201] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1196.025471] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1196.025855] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-29787770-936e-40ec-88de-268c7898d84e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.032698] env[65522]: DEBUG nova.network.neutron [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance_info_cache with network_info: [{"id": "4a2d504c-073b-4b0a-b208-065c99308adc", "address": "fa:16:3e:81:96:f8", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a2d504c-07", "ovs_interfaceid": "4a2d504c-073b-4b0a-b208-065c99308adc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1196.105156] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1196.106666] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1196.106666] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleting the datastore file [datastore2] cb62df50-2f86-469a-b2ee-fd3754d61c83 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1196.106666] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-236199df-8220-4fe1-965d-6f812f399de8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.115042] env[65522]: DEBUG oslo_vmware.api [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1196.115042] env[65522]: value = "task-5115001" [ 1196.115042] env[65522]: _type = "Task" [ 1196.115042] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.124884] env[65522]: DEBUG oslo_vmware.api [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115001, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.290139] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114999, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.307944] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 8da2a500-6f0e-4eda-9a92-79510753b3b8] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1196.406718] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52325c92-b6b6-2a50-348e-901cf948bfd0, 'name': SearchDatastore_Task, 'duration_secs': 0.017402} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.407353] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1196.407708] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1196.408276] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.408485] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1196.408762] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1196.409149] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bedc9dfb-8968-4f35-8c48-3455236daaac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.420303] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1196.420509] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1196.421509] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23da5144-a9d2-41f2-b188-705e1c9b201b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.429018] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1196.429018] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bb284e-5b87-758a-3477-e72ca774a176" [ 1196.429018] env[65522]: _type = "Task" [ 1196.429018] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.440831] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bb284e-5b87-758a-3477-e72ca774a176, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.536398] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1196.537141] env[65522]: DEBUG nova.compute.manager [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Instance network_info: |[{"id": "4a2d504c-073b-4b0a-b208-065c99308adc", "address": "fa:16:3e:81:96:f8", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a2d504c-07", "ovs_interfaceid": "4a2d504c-073b-4b0a-b208-065c99308adc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1196.537373] env[65522]: DEBUG oslo_concurrency.lockutils [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] Acquired lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1196.537620] env[65522]: DEBUG nova.network.neutron [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Refreshing network info cache for port 4a2d504c-073b-4b0a-b208-065c99308adc {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1196.539304] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:96:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a91c3a96-63d0-407c-bcde-c3d5b58d9cb2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4a2d504c-073b-4b0a-b208-065c99308adc', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1196.549533] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1196.551102] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1196.551995] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c297949-c02a-45bf-b551-46aaedc50fef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.576783] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1196.576783] env[65522]: value = "task-5115002" [ 1196.576783] env[65522]: _type = "Task" [ 1196.576783] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.588083] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115002, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.629189] env[65522]: DEBUG oslo_vmware.api [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115001, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.268376} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.629471] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1196.629654] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1196.629828] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1196.630011] env[65522]: INFO nova.compute.manager [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Took 2.14 seconds to destroy the instance on the hypervisor. [ 1196.630261] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1196.630463] env[65522]: DEBUG nova.compute.manager [-] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1196.630567] env[65522]: DEBUG nova.network.neutron [-] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1196.631786] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1196.631786] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1196.631786] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.728661] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1196.791177] env[65522]: DEBUG oslo_vmware.api [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5114999, 'name': PowerOnVM_Task, 'duration_secs': 0.658196} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.791548] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1196.791772] env[65522]: INFO nova.compute.manager [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Took 7.83 seconds to spawn the instance on the hypervisor. [ 1196.792047] env[65522]: DEBUG nova.compute.manager [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1196.793092] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ebdcfba-9fe0-480b-a1bf-cd8a4e420519 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.811373] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 84d4c689-b437-41a5-bb62-d144a30c6219] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1196.941549] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bb284e-5b87-758a-3477-e72ca774a176, 'name': SearchDatastore_Task, 'duration_secs': 0.022409} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.942393] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c0e72d9-05f0-452f-af92-d93fbf42ae0d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.950876] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1196.950876] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52615492-7c59-cf3f-d6a9-0d1ed729ce39" [ 1196.950876] env[65522]: _type = "Task" [ 1196.950876] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.962694] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52615492-7c59-cf3f-d6a9-0d1ed729ce39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.051662] env[65522]: WARNING neutronclient.v2_0.client [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1197.052449] env[65522]: WARNING openstack [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1197.053776] env[65522]: WARNING openstack [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1197.090242] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115002, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.185963] env[65522]: WARNING openstack [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1197.186860] env[65522]: WARNING openstack [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1197.241828] env[65522]: DEBUG nova.compute.manager [req-70c072c8-8105-42cd-a080-4e2b3bc1d099 req-b6600e36-30ac-40df-94f9-95ad868ee0c6 service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Received event network-vif-deleted-a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1197.242096] env[65522]: INFO nova.compute.manager [req-70c072c8-8105-42cd-a080-4e2b3bc1d099 req-b6600e36-30ac-40df-94f9-95ad868ee0c6 service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Neutron deleted interface a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac; detaching it from the instance and deleting it from the info cache [ 1197.242223] env[65522]: DEBUG nova.network.neutron [req-70c072c8-8105-42cd-a080-4e2b3bc1d099 req-b6600e36-30ac-40df-94f9-95ad868ee0c6 service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1197.312655] env[65522]: INFO nova.compute.manager [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Took 19.01 seconds to build instance. [ 1197.315459] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 9fb8b9ae-d121-4ca1-a5e2-555ddcbf0ba2] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1197.465736] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52615492-7c59-cf3f-d6a9-0d1ed729ce39, 'name': SearchDatastore_Task, 'duration_secs': 0.018139} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.466062] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1197.466358] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 874b6a07-ebba-4f75-937e-2281af1598b8/874b6a07-ebba-4f75-937e-2281af1598b8.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1197.466664] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-629c5421-8692-4218-a2c9-827d085d6495 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.479526] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1197.479526] env[65522]: value = "task-5115004" [ 1197.479526] env[65522]: _type = "Task" [ 1197.479526] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.491146] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115004, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.594430] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115002, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.704838] env[65522]: DEBUG nova.network.neutron [-] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1197.720501] env[65522]: WARNING neutronclient.v2_0.client [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1197.721245] env[65522]: WARNING openstack [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1197.721632] env[65522]: WARNING openstack [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1197.745779] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b4bc110c-2f8b-4c5c-8375-0d7ec001e06c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.764023] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee674b0c-08c5-44cf-b752-1d4e219a906b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.803907] env[65522]: DEBUG nova.compute.manager [req-70c072c8-8105-42cd-a080-4e2b3bc1d099 req-b6600e36-30ac-40df-94f9-95ad868ee0c6 service nova] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Detach interface failed, port_id=a7fc03fa-06b2-4f50-aeb7-cfc0aaf6bcac, reason: Instance cb62df50-2f86-469a-b2ee-fd3754d61c83 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1197.815284] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c148619a-c77b-435e-9e59-989ad4158b05 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.519s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1197.818966] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 21715aa5-24d4-423b-92a0-be7cc1f60877] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1197.850461] env[65522]: DEBUG nova.network.neutron [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updated VIF entry in instance network info cache for port 4a2d504c-073b-4b0a-b208-065c99308adc. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1197.851054] env[65522]: DEBUG nova.network.neutron [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance_info_cache with network_info: [{"id": "4a2d504c-073b-4b0a-b208-065c99308adc", "address": "fa:16:3e:81:96:f8", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a2d504c-07", "ovs_interfaceid": "4a2d504c-073b-4b0a-b208-065c99308adc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1197.992711] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115004, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.089343] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115002, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.207658] env[65522]: INFO nova.compute.manager [-] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Took 1.58 seconds to deallocate network for instance. [ 1198.323014] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 2a12a0a3-913e-4ade-85ef-cca7ba508ac9] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1198.355121] env[65522]: DEBUG oslo_concurrency.lockutils [req-5e83711d-adc5-4c3c-a20c-08f5e796c60b req-93dfa408-cda5-418b-b51d-1a33193bac48 service nova] Releasing lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1198.493765] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115004, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.589807] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115002, 'name': CreateVM_Task, 'duration_secs': 1.53738} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.589995] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1198.590535] env[65522]: WARNING neutronclient.v2_0.client [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1198.590916] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1198.591078] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1198.591457] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1198.591733] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d25e853-faf1-4e91-87a4-b18180f52e4e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.597428] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1198.597428] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b0940c-0149-6aaf-a14d-a244ce2cbd19" [ 1198.597428] env[65522]: _type = "Task" [ 1198.597428] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.606551] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b0940c-0149-6aaf-a14d-a244ce2cbd19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.714849] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1198.715171] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1198.715425] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1198.735234] env[65522]: INFO nova.scheduler.client.report [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleted allocations for instance cb62df50-2f86-469a-b2ee-fd3754d61c83 [ 1198.828636] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 5ce4a286-efd1-4bbc-a23b-931c6701cfe4] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1198.990930] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115004, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.147521} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.991234] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 874b6a07-ebba-4f75-937e-2281af1598b8/874b6a07-ebba-4f75-937e-2281af1598b8.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1198.991440] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1198.991714] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71fd15a5-7415-4789-95c7-03d2bceba1df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.000600] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1199.000600] env[65522]: value = "task-5115005" [ 1199.000600] env[65522]: _type = "Task" [ 1199.000600] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.011587] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115005, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.108469] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b0940c-0149-6aaf-a14d-a244ce2cbd19, 'name': SearchDatastore_Task, 'duration_secs': 0.017919} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.108805] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1199.109060] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1199.109307] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.109761] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1199.109761] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1199.110048] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70dd4d1d-65ed-4e74-b8e9-3a04564e2e12 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.120975] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1199.121162] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1199.123733] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d89bfeb-142f-4305-8b83-c9123b6c1e4d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.129580] env[65522]: DEBUG nova.compute.manager [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Stashing vm_state: active {{(pid=65522) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1199.134587] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1199.134587] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c0cce9-5615-0a26-5a78-3942399df7ef" [ 1199.134587] env[65522]: _type = "Task" [ 1199.134587] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.145232] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c0cce9-5615-0a26-5a78-3942399df7ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.243925] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e437122d-1a9e-4c2b-95a1-063b8edf418e tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "cb62df50-2f86-469a-b2ee-fd3754d61c83" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.278s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1199.332403] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: a2a5fb2d-62e0-4809-a01f-f1df66dca58f] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1199.521049] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115005, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.10407} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.521327] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1199.522123] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-702068b0-7889-4bb4-a0af-5839fadaebca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.548846] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 874b6a07-ebba-4f75-937e-2281af1598b8/874b6a07-ebba-4f75-937e-2281af1598b8.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1199.549218] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce518db1-c037-4b45-8354-dd548d27fd8a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.571176] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1199.571176] env[65522]: value = "task-5115006" [ 1199.571176] env[65522]: _type = "Task" [ 1199.571176] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.581160] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115006, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.649881] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c0cce9-5615-0a26-5a78-3942399df7ef, 'name': SearchDatastore_Task, 'duration_secs': 0.016432} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.650835] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65e5b5f9-8e1f-4fd0-8e38-bd678f58b8e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.655229] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1199.655515] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1199.658327] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1199.658327] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528752f0-2af4-cfa0-f5c2-093bf8f18c06" [ 1199.658327] env[65522]: _type = "Task" [ 1199.658327] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.668951] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528752f0-2af4-cfa0-f5c2-093bf8f18c06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.836318] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 9c39b722-56ff-44fa-8f66-3e3432645a68] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1200.085800] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115006, 'name': ReconfigVM_Task, 'duration_secs': 0.452338} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.086282] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 874b6a07-ebba-4f75-937e-2281af1598b8/874b6a07-ebba-4f75-937e-2281af1598b8.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1200.087205] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-253db7e1-6586-452a-bdd2-cb7eee051d28 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.097359] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1200.097359] env[65522]: value = "task-5115007" [ 1200.097359] env[65522]: _type = "Task" [ 1200.097359] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.113920] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115007, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.162958] env[65522]: INFO nova.compute.claims [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1200.177869] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]528752f0-2af4-cfa0-f5c2-093bf8f18c06, 'name': SearchDatastore_Task, 'duration_secs': 0.019167} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.178199] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1200.178534] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 3adfefa5-fb19-44b2-b3c5-42f2e2918673/3adfefa5-fb19-44b2-b3c5-42f2e2918673.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1200.178846] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c0b79b8c-abb4-45d3-964b-7350f57ebce3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.187169] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1200.187169] env[65522]: value = "task-5115008" [ 1200.187169] env[65522]: _type = "Task" [ 1200.187169] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.198451] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115008, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.340395] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 55ab4771-9908-4640-a142-3cb40c0c6ee5] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1200.613224] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115007, 'name': Rename_Task, 'duration_secs': 0.213371} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.613631] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1200.613936] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4081d500-986e-48ec-a72a-3cb36129826c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.626697] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1200.626697] env[65522]: value = "task-5115009" [ 1200.626697] env[65522]: _type = "Task" [ 1200.626697] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.640281] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115009, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.672681] env[65522]: INFO nova.compute.resource_tracker [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating resource usage from migration afd6f5d0-9391-4e8c-9e6f-16d7216ef2c5 [ 1200.702873] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115008, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.844207] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 9da312a0-c6eb-4903-a7d8-e178aa54d2ea] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1200.846867] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63142bca-756e-45cf-bee7-c8e0053febc0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.857621] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8459ab-5d23-4cf0-983f-138a471ae673 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.891484] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c56780-58f9-4ede-84f0-fdeb611655a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.901187] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9338db2b-237a-4b20-a385-041d92e7cd90 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.916992] env[65522]: DEBUG nova.compute.provider_tree [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1201.138350] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115009, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.199493] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115008, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.669779} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.199877] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 3adfefa5-fb19-44b2-b3c5-42f2e2918673/3adfefa5-fb19-44b2-b3c5-42f2e2918673.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1201.200155] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1201.200467] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5630bac9-f4fa-4be7-b646-69b1d0c7624d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.209167] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1201.209167] env[65522]: value = "task-5115010" [ 1201.209167] env[65522]: _type = "Task" [ 1201.209167] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.221103] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115010, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.351379] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 4264998b-949d-4b01-8832-fb1df6214403] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1201.420778] env[65522]: DEBUG nova.scheduler.client.report [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1201.521496] env[65522]: DEBUG oslo_concurrency.lockutils [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "daffdd40-9ee3-4a80-b670-5772395a32ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1201.522220] env[65522]: DEBUG oslo_concurrency.lockutils [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "daffdd40-9ee3-4a80-b670-5772395a32ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1201.637637] env[65522]: DEBUG oslo_vmware.api [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115009, 'name': PowerOnVM_Task, 'duration_secs': 0.641043} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.638155] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1201.638381] env[65522]: INFO nova.compute.manager [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Took 10.93 seconds to spawn the instance on the hypervisor. [ 1201.638564] env[65522]: DEBUG nova.compute.manager [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1201.639384] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29e3976-c8e8-47ed-a311-9f3c5fbdc2e0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.721079] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115010, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083722} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.721400] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1201.722313] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc423c6f-9258-483f-a3a2-24d7077dfac1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.746152] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] 3adfefa5-fb19-44b2-b3c5-42f2e2918673/3adfefa5-fb19-44b2-b3c5-42f2e2918673.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1201.746550] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7adc5b6-d84b-4429-9a22-328f58f82808 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.768977] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1201.768977] env[65522]: value = "task-5115011" [ 1201.768977] env[65522]: _type = "Task" [ 1201.768977] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.778606] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115011, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.855493] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 308330f8-1219-4c02-a129-22b29f0e33b2] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1201.926410] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.271s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1201.926623] env[65522]: INFO nova.compute.manager [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Migrating [ 1202.026187] env[65522]: DEBUG nova.compute.manager [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1202.157215] env[65522]: INFO nova.compute.manager [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Took 22.44 seconds to build instance. [ 1202.166825] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d9517-44f9-afc3-b6a1-d09bd16ad6fa/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1202.167883] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf4d0c3f-7011-4ef8-9d80-0d3b1b1326dc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.175610] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d9517-44f9-afc3-b6a1-d09bd16ad6fa/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1202.175818] env[65522]: ERROR oslo_vmware.rw_handles [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d9517-44f9-afc3-b6a1-d09bd16ad6fa/disk-0.vmdk due to incomplete transfer. [ 1202.176041] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1a314182-28d5-46d4-b709-ae65547595fb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.184346] env[65522]: DEBUG oslo_vmware.rw_handles [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521d9517-44f9-afc3-b6a1-d09bd16ad6fa/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1202.184600] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Uploaded image 7832ad6e-429a-4894-9720-58fd3520c4e2 to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1202.187278] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1202.187580] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4378319e-6aef-4995-b27e-a3a8c9296188 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.197061] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1202.197061] env[65522]: value = "task-5115012" [ 1202.197061] env[65522]: _type = "Task" [ 1202.197061] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.207811] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115012, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.281686] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115011, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.360424] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 1e7b8237-34ea-479e-b5a7-b1846661d61d] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1202.443438] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1202.443659] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1202.443852] env[65522]: DEBUG nova.network.neutron [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1202.554787] env[65522]: DEBUG oslo_concurrency.lockutils [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.555330] env[65522]: DEBUG oslo_concurrency.lockutils [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1202.557044] env[65522]: INFO nova.compute.claims [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1202.661450] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e6440f08-e4be-4977-b19b-f8d414ea2a03 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "874b6a07-ebba-4f75-937e-2281af1598b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.966s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1202.709384] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115012, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.780964] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115011, 'name': ReconfigVM_Task, 'duration_secs': 0.793057} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.781333] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Reconfigured VM instance instance-00000073 to attach disk [datastore1] 3adfefa5-fb19-44b2-b3c5-42f2e2918673/3adfefa5-fb19-44b2-b3c5-42f2e2918673.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1202.782012] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-47268252-8dc4-422a-b884-8719609feaeb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.789898] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1202.789898] env[65522]: value = "task-5115013" [ 1202.789898] env[65522]: _type = "Task" [ 1202.789898] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.798713] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115013, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.862758] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 9df1a51a-2811-4486-a4c6-58d618f2ae7d] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1202.947562] env[65522]: WARNING neutronclient.v2_0.client [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1202.948417] env[65522]: WARNING openstack [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1202.949014] env[65522]: WARNING openstack [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1203.050044] env[65522]: WARNING openstack [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1203.050674] env[65522]: WARNING openstack [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1203.119045] env[65522]: WARNING neutronclient.v2_0.client [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1203.119784] env[65522]: WARNING openstack [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1203.120099] env[65522]: WARNING openstack [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1203.146668] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1203.146906] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1203.191029] env[65522]: INFO nova.compute.manager [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Rescuing [ 1203.191276] env[65522]: DEBUG oslo_concurrency.lockutils [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "refresh_cache-874b6a07-ebba-4f75-937e-2281af1598b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1203.191461] env[65522]: DEBUG oslo_concurrency.lockutils [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquired lock "refresh_cache-874b6a07-ebba-4f75-937e-2281af1598b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1203.191642] env[65522]: DEBUG nova.network.neutron [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1203.208913] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115012, 'name': Destroy_Task, 'duration_secs': 0.967826} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.211328] env[65522]: DEBUG nova.network.neutron [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating instance_info_cache with network_info: [{"id": "499e91eb-948a-4b54-b7a6-ce30af8315ea", "address": "fa:16:3e:09:07:e8", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap499e91eb-94", "ovs_interfaceid": "499e91eb-948a-4b54-b7a6-ce30af8315ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1203.212471] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Destroyed the VM [ 1203.212747] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1203.213843] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8e98e7fa-c8eb-4c94-bee5-a909ee607c76 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.222037] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1203.222037] env[65522]: value = "task-5115014" [ 1203.222037] env[65522]: _type = "Task" [ 1203.222037] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.231730] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115014, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.300721] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115013, 'name': Rename_Task, 'duration_secs': 0.324776} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.301031] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1203.301296] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98388fb4-cd28-40aa-a46e-8e2a876e8481 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.308773] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1203.308773] env[65522]: value = "task-5115015" [ 1203.308773] env[65522]: _type = "Task" [ 1203.308773] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.318177] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115015, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.366632] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 6ee4f79a-b6f2-4f8c-b049-d00b38ac4a8e] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1203.427670] env[65522]: DEBUG oslo_vmware.rw_handles [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522d7da7-f0e0-d089-e742-411604bc1ffb/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1203.428845] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dddf8da4-daa3-4d74-8b10-249c4a6881be {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.436475] env[65522]: DEBUG oslo_vmware.rw_handles [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522d7da7-f0e0-d089-e742-411604bc1ffb/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1203.436683] env[65522]: ERROR oslo_vmware.rw_handles [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522d7da7-f0e0-d089-e742-411604bc1ffb/disk-0.vmdk due to incomplete transfer. [ 1203.437053] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9b4d2002-bfe5-4830-be36-b85f3e9af636 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.447679] env[65522]: DEBUG oslo_vmware.rw_handles [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522d7da7-f0e0-d089-e742-411604bc1ffb/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1203.447679] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Uploaded image e4e8fd5f-0bc1-437d-a8db-457c5ea05344 to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1203.448826] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1203.449222] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6990557f-2c45-4329-9ded-dfdd6a845519 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.458167] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1203.458167] env[65522]: value = "task-5115016" [ 1203.458167] env[65522]: _type = "Task" [ 1203.458167] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.471922] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115016, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.652280] env[65522]: DEBUG nova.compute.manager [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1203.695031] env[65522]: WARNING neutronclient.v2_0.client [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1203.695813] env[65522]: WARNING openstack [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1203.696405] env[65522]: WARNING openstack [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1203.714978] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1203.731609] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221a3994-2abb-4ab7-862d-bff81f603dd9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.737066] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115014, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.747038] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df779ec7-9947-420f-8e90-3be1536a2cf8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.782258] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a159a0af-e240-4f4a-b6f7-2bc091ece2c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.790767] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e781b5d-3d08-409e-b79d-9cc838db73a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.811529] env[65522]: DEBUG nova.compute.provider_tree [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1203.818037] env[65522]: WARNING openstack [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1203.818504] env[65522]: WARNING openstack [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1203.833472] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115015, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.870631] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 179a37ed-1176-4082-8c68-1be6abd3fd9d] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1203.893459] env[65522]: WARNING neutronclient.v2_0.client [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1203.894608] env[65522]: WARNING openstack [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1203.894841] env[65522]: WARNING openstack [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1203.969033] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115016, 'name': Destroy_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.003407] env[65522]: DEBUG nova.network.neutron [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Updating instance_info_cache with network_info: [{"id": "54d3b471-cec2-4c58-a51d-7bd3ed42e50e", "address": "fa:16:3e:08:2f:12", "network": {"id": "26dda233-0dd0-40ae-b56f-c84b1b91e92a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1197499239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "15229af4da7a4ff9963369df50fb9d5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap54d3b471-ce", "ovs_interfaceid": "54d3b471-cec2-4c58-a51d-7bd3ed42e50e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1204.176069] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1204.233832] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115014, 'name': RemoveSnapshot_Task, 'duration_secs': 0.750963} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.234114] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1204.234389] env[65522]: DEBUG nova.compute.manager [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1204.235188] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1788087-2e42-4432-bd77-b587a583c27b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.320566] env[65522]: DEBUG oslo_vmware.api [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115015, 'name': PowerOnVM_Task, 'duration_secs': 0.764087} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.320866] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1204.321082] env[65522]: INFO nova.compute.manager [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Took 10.81 seconds to spawn the instance on the hypervisor. [ 1204.321285] env[65522]: DEBUG nova.compute.manager [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1204.322077] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dddde33-4e4a-47f4-a34a-9b6379b47574 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.328842] env[65522]: DEBUG nova.scheduler.client.report [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1204.374525] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 8eda6816-6781-4ad9-9210-11fa623f24b0] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1204.470931] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115016, 'name': Destroy_Task, 'duration_secs': 0.539357} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.471312] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Destroyed the VM [ 1204.471587] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1204.471876] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d5bace72-1e4f-42c4-b95d-3e8b90fcd90a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.478833] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1204.478833] env[65522]: value = "task-5115017" [ 1204.478833] env[65522]: _type = "Task" [ 1204.478833] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.490103] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115017, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.506178] env[65522]: DEBUG oslo_concurrency.lockutils [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Releasing lock "refresh_cache-874b6a07-ebba-4f75-937e-2281af1598b8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1204.747492] env[65522]: INFO nova.compute.manager [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Shelve offloading [ 1204.836252] env[65522]: DEBUG oslo_concurrency.lockutils [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.281s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1204.836840] env[65522]: DEBUG nova.compute.manager [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1204.844696] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.669s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1204.846176] env[65522]: INFO nova.compute.claims [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1204.851326] env[65522]: INFO nova.compute.manager [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Took 20.90 seconds to build instance. [ 1204.877763] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 5e227f6c-9f13-416f-8e6d-2f7d931619fd] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1204.990683] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115017, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.216878] env[65522]: DEBUG nova.compute.manager [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Received event network-changed-4a2d504c-073b-4b0a-b208-065c99308adc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1205.216979] env[65522]: DEBUG nova.compute.manager [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Refreshing instance network info cache due to event network-changed-4a2d504c-073b-4b0a-b208-065c99308adc. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1205.217209] env[65522]: DEBUG oslo_concurrency.lockutils [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] Acquiring lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1205.217339] env[65522]: DEBUG oslo_concurrency.lockutils [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] Acquired lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1205.217496] env[65522]: DEBUG nova.network.neutron [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Refreshing network info cache for port 4a2d504c-073b-4b0a-b208-065c99308adc {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1205.232329] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13c734fc-47ba-4861-b6f2-64197054c00c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.252128] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1205.252497] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating instance 'ae3455d8-2f38-42ad-b16d-d98aef92b2a8' progress to 0 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1205.256112] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43b1ea77-2c63-4ce0-ab23-39b5ed70f242 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.263561] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1205.263561] env[65522]: value = "task-5115018" [ 1205.263561] env[65522]: _type = "Task" [ 1205.263561] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.273023] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115018, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.347415] env[65522]: DEBUG nova.compute.utils [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1205.348823] env[65522]: DEBUG nova.compute.manager [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1205.349030] env[65522]: DEBUG nova.network.neutron [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1205.349379] env[65522]: WARNING neutronclient.v2_0.client [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1205.349768] env[65522]: WARNING neutronclient.v2_0.client [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1205.350372] env[65522]: WARNING openstack [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1205.350717] env[65522]: WARNING openstack [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1205.360040] env[65522]: DEBUG oslo_concurrency.lockutils [None req-27ca7969-6a1e-4760-a46d-eb38cb701265 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.420s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1205.382088] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 26432ad6-610e-49c2-8699-f1e7e128e5b8] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1205.399932] env[65522]: DEBUG nova.policy [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f010fe7784043aaa3d4d44b921c438f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91597233ae9c44c094f4c32d90332fa6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1205.490808] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115017, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.721454] env[65522]: WARNING neutronclient.v2_0.client [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1205.722255] env[65522]: WARNING openstack [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1205.722631] env[65522]: WARNING openstack [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1205.760578] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1205.760937] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00b659cc-b9e3-4c31-89ad-4d490dbad07b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.764175] env[65522]: DEBUG nova.network.neutron [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Successfully created port: 6b80135d-6228-4366-83aa-ff70808ae005 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1205.771662] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1205.771662] env[65522]: value = "task-5115019" [ 1205.771662] env[65522]: _type = "Task" [ 1205.771662] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.779381] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1205.779620] env[65522]: DEBUG nova.compute.manager [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1205.780448] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7caa3acb-35de-4c21-9134-44b885736066 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.786037] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115019, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.790377] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1205.790580] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquired lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1205.790712] env[65522]: DEBUG nova.network.neutron [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1205.858544] env[65522]: DEBUG nova.compute.manager [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1205.887768] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 77b3ca48-d2c6-4743-a44a-53b9aa84662a] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1205.976354] env[65522]: WARNING openstack [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1205.976940] env[65522]: WARNING openstack [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1205.996741] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115017, 'name': RemoveSnapshot_Task, 'duration_secs': 1.162042} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.997037] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1205.997328] env[65522]: DEBUG nova.compute.manager [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1205.998300] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed1ab3a2-4939-4415-aeec-c35e6e31d199 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.048694] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1206.049035] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5323b33-e68e-4582-882d-58a2f2ca9ccd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.062031] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1206.062031] env[65522]: value = "task-5115020" [ 1206.062031] env[65522]: _type = "Task" [ 1206.062031] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.071982] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115020, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.075827] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa86224-10b2-4ed1-9339-733541976f84 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.085061] env[65522]: WARNING neutronclient.v2_0.client [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1206.085695] env[65522]: WARNING openstack [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1206.086312] env[65522]: WARNING openstack [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1206.095733] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b62043-9f71-4e81-86b9-6762714b962c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.137456] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514571a9-4656-4f3c-9275-7678a71d31e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.147241] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030cfcb7-45a9-47cc-893f-14622dc0fb9c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.162746] env[65522]: DEBUG nova.compute.provider_tree [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1206.286025] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115019, 'name': PowerOffVM_Task, 'duration_secs': 0.398464} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.286025] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1206.286025] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating instance 'ae3455d8-2f38-42ad-b16d-d98aef92b2a8' progress to 17 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1206.292605] env[65522]: DEBUG nova.network.neutron [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updated VIF entry in instance network info cache for port 4a2d504c-073b-4b0a-b208-065c99308adc. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1206.292963] env[65522]: DEBUG nova.network.neutron [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance_info_cache with network_info: [{"id": "4a2d504c-073b-4b0a-b208-065c99308adc", "address": "fa:16:3e:81:96:f8", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a2d504c-07", "ovs_interfaceid": "4a2d504c-073b-4b0a-b208-065c99308adc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1206.298018] env[65522]: WARNING neutronclient.v2_0.client [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1206.298018] env[65522]: WARNING openstack [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1206.298018] env[65522]: WARNING openstack [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1206.366799] env[65522]: INFO nova.virt.block_device [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Booting with volume eacd0d18-4608-43ee-a54c-6655db952ccb at /dev/sda [ 1206.394103] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 0b48b747-149c-4163-bdd0-7d0b07ddacfd] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1206.408418] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c3ccdc9-418e-41f9-b2d8-75d687b9ec50 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.418517] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6a059e-897b-4bc6-aded-70c052d2bbae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.455535] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e77cf7c3-1d69-43c1-a790-d621527e418e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.464990] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780ff85d-f757-4c38-92fd-bc9a39364492 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.499223] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c644e785-3160-41c1-a235-d888ae324d8e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.506756] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0193a8d9-8f18-4425-85be-d8bbf0de0f0d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.512492] env[65522]: WARNING openstack [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1206.512874] env[65522]: WARNING openstack [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1206.527368] env[65522]: DEBUG nova.virt.block_device [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating existing volume attachment record: 2d03c130-1c0c-4fe8-96b0-05438167c3d4 {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1206.529864] env[65522]: INFO nova.compute.manager [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Shelve offloading [ 1206.575378] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115020, 'name': PowerOffVM_Task, 'duration_secs': 0.202704} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.575658] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1206.576535] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b33f1d-dab4-4e05-bad5-ead35c66c787 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.596781] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b08085-f318-4e37-a96d-4ff59cbd5ed1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.607687] env[65522]: WARNING neutronclient.v2_0.client [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1206.608409] env[65522]: WARNING openstack [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1206.608755] env[65522]: WARNING openstack [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1206.631066] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1206.631412] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea4cb8ae-2761-4667-8bc4-ed269874dfdd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.638303] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1206.638303] env[65522]: value = "task-5115021" [ 1206.638303] env[65522]: _type = "Task" [ 1206.638303] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.649301] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1206.649525] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1206.649796] env[65522]: DEBUG oslo_concurrency.lockutils [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1206.649958] env[65522]: DEBUG oslo_concurrency.lockutils [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1206.650157] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1206.650442] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a337a35e-3a64-4623-9d3b-4865998743c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.660021] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1206.660173] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1206.660953] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-222a9940-2649-4531-9db7-bedbc986b3bb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.669037] env[65522]: DEBUG nova.scheduler.client.report [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1206.674294] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1206.674294] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b1a6a4-2931-559a-ee20-b6fbe33989b6" [ 1206.674294] env[65522]: _type = "Task" [ 1206.674294] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.684532] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b1a6a4-2931-559a-ee20-b6fbe33989b6, 'name': SearchDatastore_Task, 'duration_secs': 0.010369} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.685416] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f4393c9-ada3-431f-8fd0-3d50fdb5663b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.692230] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1206.692230] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c881c8-77d2-00e8-f200-52546d7b9936" [ 1206.692230] env[65522]: _type = "Task" [ 1206.692230] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.703455] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c881c8-77d2-00e8-f200-52546d7b9936, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.712707] env[65522]: DEBUG nova.network.neutron [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Updating instance_info_cache with network_info: [{"id": "0c495486-a850-4368-95a9-03fac8060924", "address": "fa:16:3e:a7:35:38", "network": {"id": "2eb73cd0-541f-42e9-8774-600c5c1e9200", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-939435024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdba2dbc660145b7a98af8bd3f77a035", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c495486-a8", "ovs_interfaceid": "0c495486-a850-4368-95a9-03fac8060924", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1206.793233] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1206.793634] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1206.793877] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1206.794178] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1206.794403] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1206.794617] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1206.794915] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1206.795216] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1206.795510] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1206.795736] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1206.796054] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1206.804979] env[65522]: DEBUG oslo_concurrency.lockutils [req-74c25d5a-d398-4f29-bc81-40b16f823d6a req-4f1d7b27-a56a-48eb-a58a-8b83bd8f334d service nova] Releasing lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1206.805454] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c59d6aea-bbf6-449d-9062-8ddd05807030 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.828719] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1206.828719] env[65522]: value = "task-5115022" [ 1206.828719] env[65522]: _type = "Task" [ 1206.828719] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.838889] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115022, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.898729] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 18568baa-83de-4b54-addf-dee66211c8d7] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1207.034800] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1207.035192] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-683f0c5a-21d3-49a4-877c-ff70ad8429ff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.043647] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1207.043647] env[65522]: value = "task-5115023" [ 1207.043647] env[65522]: _type = "Task" [ 1207.043647] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.054375] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1207.054581] env[65522]: DEBUG nova.compute.manager [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1207.055366] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03da44bb-d916-42ec-9930-388783b67916 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.062661] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1207.062935] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1207.063186] env[65522]: DEBUG nova.network.neutron [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1207.177188] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.332s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1207.177803] env[65522]: DEBUG nova.compute.manager [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1207.204643] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c881c8-77d2-00e8-f200-52546d7b9936, 'name': SearchDatastore_Task, 'duration_secs': 0.013029} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.204936] env[65522]: DEBUG oslo_concurrency.lockutils [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1207.205223] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 874b6a07-ebba-4f75-937e-2281af1598b8/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk. {{(pid=65522) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1207.205536] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc1ebe33-2304-483c-b652-15f0178b7ba7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.212994] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1207.212994] env[65522]: value = "task-5115024" [ 1207.212994] env[65522]: _type = "Task" [ 1207.212994] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.214737] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Releasing lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1207.215289] env[65522]: WARNING neutronclient.v2_0.client [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1207.216090] env[65522]: WARNING openstack [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1207.216476] env[65522]: WARNING openstack [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1207.222918] env[65522]: WARNING neutronclient.v2_0.client [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1207.236894] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115024, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.266912] env[65522]: DEBUG nova.compute.manager [req-943f2cc2-6648-44c7-a146-1cdaa3641add req-7e87bab7-af37-43af-9cd3-0b5b33f3222e service nova] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Received event network-vif-plugged-6b80135d-6228-4366-83aa-ff70808ae005 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1207.267177] env[65522]: DEBUG oslo_concurrency.lockutils [req-943f2cc2-6648-44c7-a146-1cdaa3641add req-7e87bab7-af37-43af-9cd3-0b5b33f3222e service nova] Acquiring lock "daffdd40-9ee3-4a80-b670-5772395a32ae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1207.267342] env[65522]: DEBUG oslo_concurrency.lockutils [req-943f2cc2-6648-44c7-a146-1cdaa3641add req-7e87bab7-af37-43af-9cd3-0b5b33f3222e service nova] Lock "daffdd40-9ee3-4a80-b670-5772395a32ae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1207.267516] env[65522]: DEBUG oslo_concurrency.lockutils [req-943f2cc2-6648-44c7-a146-1cdaa3641add req-7e87bab7-af37-43af-9cd3-0b5b33f3222e service nova] Lock "daffdd40-9ee3-4a80-b670-5772395a32ae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1207.267676] env[65522]: DEBUG nova.compute.manager [req-943f2cc2-6648-44c7-a146-1cdaa3641add req-7e87bab7-af37-43af-9cd3-0b5b33f3222e service nova] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] No waiting events found dispatching network-vif-plugged-6b80135d-6228-4366-83aa-ff70808ae005 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1207.267830] env[65522]: WARNING nova.compute.manager [req-943f2cc2-6648-44c7-a146-1cdaa3641add req-7e87bab7-af37-43af-9cd3-0b5b33f3222e service nova] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Received unexpected event network-vif-plugged-6b80135d-6228-4366-83aa-ff70808ae005 for instance with vm_state building and task_state block_device_mapping. [ 1207.340874] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115022, 'name': ReconfigVM_Task, 'duration_secs': 0.254591} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.341297] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating instance 'ae3455d8-2f38-42ad-b16d-d98aef92b2a8' progress to 33 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1207.361109] env[65522]: DEBUG nova.network.neutron [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Successfully updated port: 6b80135d-6228-4366-83aa-ff70808ae005 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1207.404830] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 05319f69-2ac7-4649-a90a-3542d87a3e82] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1207.507980] env[65522]: DEBUG nova.compute.manager [req-ec8b7aab-a3de-477a-abbe-54b1eb1b560d req-091d92d8-8324-49e0-9e31-5759b86d1fb3 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Received event network-vif-unplugged-0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1207.508233] env[65522]: DEBUG oslo_concurrency.lockutils [req-ec8b7aab-a3de-477a-abbe-54b1eb1b560d req-091d92d8-8324-49e0-9e31-5759b86d1fb3 service nova] Acquiring lock "2981b635-43c8-4bd6-9991-e6af0be82f3c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1207.508448] env[65522]: DEBUG oslo_concurrency.lockutils [req-ec8b7aab-a3de-477a-abbe-54b1eb1b560d req-091d92d8-8324-49e0-9e31-5759b86d1fb3 service nova] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1207.508688] env[65522]: DEBUG oslo_concurrency.lockutils [req-ec8b7aab-a3de-477a-abbe-54b1eb1b560d req-091d92d8-8324-49e0-9e31-5759b86d1fb3 service nova] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1207.509597] env[65522]: DEBUG nova.compute.manager [req-ec8b7aab-a3de-477a-abbe-54b1eb1b560d req-091d92d8-8324-49e0-9e31-5759b86d1fb3 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] No waiting events found dispatching network-vif-unplugged-0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1207.509920] env[65522]: WARNING nova.compute.manager [req-ec8b7aab-a3de-477a-abbe-54b1eb1b560d req-091d92d8-8324-49e0-9e31-5759b86d1fb3 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Received unexpected event network-vif-unplugged-0c495486-a850-4368-95a9-03fac8060924 for instance with vm_state shelved and task_state shelving_offloading. [ 1207.530053] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1207.531225] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325598f4-6a7d-4589-8257-9f7238755ad6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.540030] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1207.540030] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5a0d5cc3-8756-4e43-abfb-a53a3882e3ac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.566475] env[65522]: WARNING neutronclient.v2_0.client [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1207.567346] env[65522]: WARNING openstack [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1207.567844] env[65522]: WARNING openstack [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1207.628134] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1207.628386] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1207.628574] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Deleting the datastore file [datastore2] 2981b635-43c8-4bd6-9991-e6af0be82f3c {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1207.628894] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0dd0c082-83f9-4dd2-a0c3-f133abb8f1b9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.639759] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1207.639759] env[65522]: value = "task-5115026" [ 1207.639759] env[65522]: _type = "Task" [ 1207.639759] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.651423] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115026, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.683099] env[65522]: DEBUG nova.compute.utils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1207.684896] env[65522]: DEBUG nova.compute.manager [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1207.685123] env[65522]: DEBUG nova.network.neutron [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1207.685544] env[65522]: WARNING neutronclient.v2_0.client [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1207.685879] env[65522]: WARNING neutronclient.v2_0.client [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1207.686552] env[65522]: WARNING openstack [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1207.686916] env[65522]: WARNING openstack [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1207.695772] env[65522]: DEBUG nova.compute.manager [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1207.724314] env[65522]: WARNING openstack [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1207.724794] env[65522]: WARNING openstack [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1207.735897] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115024, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.739689] env[65522]: DEBUG nova.policy [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18ceaff46b314c9f9da68a8d0c6f5f2b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fa11b46d9fe144f391233e6eb9c819d7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1207.790142] env[65522]: WARNING neutronclient.v2_0.client [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1207.790855] env[65522]: WARNING openstack [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1207.791283] env[65522]: WARNING openstack [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1207.854117] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1207.854494] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1207.854734] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1207.855011] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1207.855246] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1207.855460] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1207.855781] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1207.856015] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1207.856257] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1207.856483] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1207.856774] env[65522]: DEBUG nova.virt.hardware [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1207.865784] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Reconfiguring VM instance instance-00000071 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1207.867059] env[65522]: DEBUG oslo_concurrency.lockutils [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1207.867566] env[65522]: DEBUG oslo_concurrency.lockutils [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1207.867566] env[65522]: DEBUG nova.network.neutron [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1207.869137] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec6d27b9-7645-4e39-8d44-3d8e7f72fc9c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.902618] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1207.902618] env[65522]: value = "task-5115027" [ 1207.902618] env[65522]: _type = "Task" [ 1207.902618] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.908813] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 1eebb320-30e2-4d8b-b1fd-19e31fb35b77] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1207.915029] env[65522]: DEBUG nova.network.neutron [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Updating instance_info_cache with network_info: [{"id": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "address": "fa:16:3e:c3:dd:ad", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfed73ed8-c0", "ovs_interfaceid": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1207.919880] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115027, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.097103] env[65522]: DEBUG nova.network.neutron [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Successfully created port: 0a9cf0d6-c5cb-4805-918d-ed092924fb2a {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1208.151383] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115026, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.226351] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115024, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.779013} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.226598] env[65522]: INFO nova.virt.vmwareapi.ds_util [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 874b6a07-ebba-4f75-937e-2281af1598b8/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk. [ 1208.227376] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e00a0d1-338e-4f33-a8d2-33bda74e768c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.256622] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 874b6a07-ebba-4f75-937e-2281af1598b8/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1208.257737] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f17866c-34cf-4958-b7f1-9e88464726da {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.277168] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1208.277168] env[65522]: value = "task-5115028" [ 1208.277168] env[65522]: _type = "Task" [ 1208.277168] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.286508] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115028, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.392352] env[65522]: WARNING openstack [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1208.392773] env[65522]: WARNING openstack [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1208.412705] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115027, 'name': ReconfigVM_Task, 'duration_secs': 0.265886} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.412985] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Reconfigured VM instance instance-00000071 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1208.413817] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330a2d51-f183-4b08-b596-9ef438044edb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.429242] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 495fa98f-a9db-4214-87cc-d29209d3cb62] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1208.431750] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1208.431856] env[65522]: WARNING neutronclient.v2_0.client [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1208.432510] env[65522]: WARNING openstack [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1208.432935] env[65522]: WARNING openstack [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1208.437902] env[65522]: WARNING neutronclient.v2_0.client [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1208.447150] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] ae3455d8-2f38-42ad-b16d-d98aef92b2a8/ae3455d8-2f38-42ad-b16d-d98aef92b2a8.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1208.448684] env[65522]: DEBUG nova.network.neutron [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1208.450446] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a95d26a0-7f2e-4524-bd34-ced5a96d9f2d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.469824] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1208.469824] env[65522]: value = "task-5115029" [ 1208.469824] env[65522]: _type = "Task" [ 1208.469824] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.480533] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115029, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.490180] env[65522]: WARNING openstack [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1208.490566] env[65522]: WARNING openstack [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1208.620199] env[65522]: DEBUG nova.compute.manager [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1208.620846] env[65522]: DEBUG nova.virt.hardware [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1208.621096] env[65522]: DEBUG nova.virt.hardware [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1208.621292] env[65522]: DEBUG nova.virt.hardware [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1208.621517] env[65522]: DEBUG nova.virt.hardware [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1208.621696] env[65522]: DEBUG nova.virt.hardware [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1208.621877] env[65522]: DEBUG nova.virt.hardware [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1208.622125] env[65522]: DEBUG nova.virt.hardware [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1208.622295] env[65522]: DEBUG nova.virt.hardware [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1208.622462] env[65522]: DEBUG nova.virt.hardware [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1208.622628] env[65522]: DEBUG nova.virt.hardware [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1208.622794] env[65522]: DEBUG nova.virt.hardware [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1208.624594] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0662e073-c7e5-4a7f-9480-6b34d2dbb276 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.633841] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4867a9f-caa8-476c-a14d-b7e394fda42c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.662359] env[65522]: DEBUG oslo_vmware.api [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115026, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.542764} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.662631] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1208.662818] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1208.662988] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1208.686674] env[65522]: INFO nova.scheduler.client.report [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Deleted allocations for instance 2981b635-43c8-4bd6-9991-e6af0be82f3c [ 1208.706299] env[65522]: DEBUG nova.compute.manager [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1208.737959] env[65522]: DEBUG nova.virt.hardware [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1208.738333] env[65522]: DEBUG nova.virt.hardware [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1208.738496] env[65522]: DEBUG nova.virt.hardware [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1208.738687] env[65522]: DEBUG nova.virt.hardware [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1208.738834] env[65522]: DEBUG nova.virt.hardware [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1208.739075] env[65522]: DEBUG nova.virt.hardware [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1208.739258] env[65522]: DEBUG nova.virt.hardware [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1208.739418] env[65522]: DEBUG nova.virt.hardware [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1208.739581] env[65522]: DEBUG nova.virt.hardware [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1208.739751] env[65522]: DEBUG nova.virt.hardware [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1208.739922] env[65522]: DEBUG nova.virt.hardware [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1208.741091] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b09602-f66d-4978-8ba8-8ac29bf4e959 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.749542] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7412a3fb-8c36-4009-909b-10f35f538dfd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.787584] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115028, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.948532] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 24f520ec-6a56-4f17-9ae4-2c856bf99582] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1208.962223] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1208.963173] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eaf5a38-5dde-4a14-b6f6-6010333cabd4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.971902] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1208.975062] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a9136b02-b613-4f41-b641-f2d25d3b8be7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.982896] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115029, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.014708] env[65522]: WARNING neutronclient.v2_0.client [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1209.015431] env[65522]: WARNING openstack [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1209.015904] env[65522]: WARNING openstack [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1209.086394] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1209.086779] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1209.086839] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleting the datastore file [datastore2] 3b82f9ce-ea59-430d-adb6-918cc6cc48a4 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1209.087267] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e42b2558-f7c1-4ef6-a401-449fef629b7d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.096724] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1209.096724] env[65522]: value = "task-5115031" [ 1209.096724] env[65522]: _type = "Task" [ 1209.096724] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.107890] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115031, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.191236] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1209.191829] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1209.192125] env[65522]: DEBUG nova.objects.instance [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lazy-loading 'resources' on Instance uuid 2981b635-43c8-4bd6-9991-e6af0be82f3c {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1209.199611] env[65522]: DEBUG nova.network.neutron [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating instance_info_cache with network_info: [{"id": "6b80135d-6228-4366-83aa-ff70808ae005", "address": "fa:16:3e:82:79:a0", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b80135d-62", "ovs_interfaceid": "6b80135d-6228-4366-83aa-ff70808ae005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1209.291946] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115028, 'name': ReconfigVM_Task, 'duration_secs': 0.893814} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.292513] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 874b6a07-ebba-4f75-937e-2281af1598b8/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1209.293799] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e14dc2-4c97-4e9e-b393-983813caf5c7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.301294] env[65522]: DEBUG nova.compute.manager [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Received event network-changed-6b80135d-6228-4366-83aa-ff70808ae005 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1209.302133] env[65522]: DEBUG nova.compute.manager [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Refreshing instance network info cache due to event network-changed-6b80135d-6228-4366-83aa-ff70808ae005. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1209.302493] env[65522]: DEBUG oslo_concurrency.lockutils [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] Acquiring lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1209.326273] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6eb3a388-5410-438e-8a20-628a23452975 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.344317] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1209.344317] env[65522]: value = "task-5115032" [ 1209.344317] env[65522]: _type = "Task" [ 1209.344317] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.353757] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115032, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.451933] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: bb20d99f-fc49-47bf-a6b7-74ea966b71eb] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1209.485849] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115029, 'name': ReconfigVM_Task, 'duration_secs': 0.797728} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.486170] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Reconfigured VM instance instance-00000071 to attach disk [datastore2] ae3455d8-2f38-42ad-b16d-d98aef92b2a8/ae3455d8-2f38-42ad-b16d-d98aef92b2a8.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1209.486452] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating instance 'ae3455d8-2f38-42ad-b16d-d98aef92b2a8' progress to 50 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1209.546996] env[65522]: DEBUG nova.compute.manager [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Received event network-changed-0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1209.547218] env[65522]: DEBUG nova.compute.manager [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Refreshing instance network info cache due to event network-changed-0c495486-a850-4368-95a9-03fac8060924. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1209.547442] env[65522]: DEBUG oslo_concurrency.lockutils [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Acquiring lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1209.547637] env[65522]: DEBUG oslo_concurrency.lockutils [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Acquired lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1209.547873] env[65522]: DEBUG nova.network.neutron [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Refreshing network info cache for port 0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1209.610384] env[65522]: DEBUG oslo_vmware.api [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115031, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.349495} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.610757] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1209.610953] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1209.611850] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1209.618619] env[65522]: DEBUG nova.network.neutron [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Successfully updated port: 0a9cf0d6-c5cb-4805-918d-ed092924fb2a {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1209.638654] env[65522]: INFO nova.scheduler.client.report [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleted allocations for instance 3b82f9ce-ea59-430d-adb6-918cc6cc48a4 [ 1209.695440] env[65522]: DEBUG nova.objects.instance [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lazy-loading 'numa_topology' on Instance uuid 2981b635-43c8-4bd6-9991-e6af0be82f3c {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1209.702574] env[65522]: DEBUG oslo_concurrency.lockutils [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1209.702967] env[65522]: DEBUG nova.compute.manager [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Instance network_info: |[{"id": "6b80135d-6228-4366-83aa-ff70808ae005", "address": "fa:16:3e:82:79:a0", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b80135d-62", "ovs_interfaceid": "6b80135d-6228-4366-83aa-ff70808ae005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1209.703702] env[65522]: DEBUG oslo_concurrency.lockutils [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] Acquired lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1209.703888] env[65522]: DEBUG nova.network.neutron [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Refreshing network info cache for port 6b80135d-6228-4366-83aa-ff70808ae005 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1209.705152] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:79:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56136ef6-99d7-4562-9a9f-d66fec951c5c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6b80135d-6228-4366-83aa-ff70808ae005', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1209.712807] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1209.714115] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1209.714369] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6a849289-b148-4749-baf3-07b19c873a17 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.736488] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1209.736488] env[65522]: value = "task-5115033" [ 1209.736488] env[65522]: _type = "Task" [ 1209.736488] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.745538] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115033, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.856084] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115032, 'name': ReconfigVM_Task, 'duration_secs': 0.220293} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.856260] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1209.856488] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c42c6d0-1663-4ce3-8925-edc37bda2127 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.865039] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1209.865039] env[65522]: value = "task-5115034" [ 1209.865039] env[65522]: _type = "Task" [ 1209.865039] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.874513] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115034, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.956035] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 42a08cfb-d865-4967-a086-370a2ca98b7d] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1209.993628] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9445ff-7ed7-4e56-9727-1fc4758da534 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.015524] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3611ecf6-912c-4cd6-9475-1e54b4445f05 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.036352] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating instance 'ae3455d8-2f38-42ad-b16d-d98aef92b2a8' progress to 67 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1210.050881] env[65522]: WARNING neutronclient.v2_0.client [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1210.051697] env[65522]: WARNING openstack [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1210.052095] env[65522]: WARNING openstack [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1210.122324] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.122324] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1210.122324] env[65522]: DEBUG nova.network.neutron [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1210.143679] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1210.166708] env[65522]: WARNING openstack [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1210.167119] env[65522]: WARNING openstack [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1210.200976] env[65522]: DEBUG nova.objects.base [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Object Instance<2981b635-43c8-4bd6-9991-e6af0be82f3c> lazy-loaded attributes: resources,numa_topology {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1210.214728] env[65522]: WARNING neutronclient.v2_0.client [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1210.215643] env[65522]: WARNING openstack [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1210.216108] env[65522]: WARNING openstack [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1210.235881] env[65522]: WARNING neutronclient.v2_0.client [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1210.237032] env[65522]: WARNING openstack [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1210.237032] env[65522]: WARNING openstack [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1210.263902] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115033, 'name': CreateVM_Task, 'duration_secs': 0.363095} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.266590] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1210.268327] env[65522]: WARNING neutronclient.v2_0.client [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1210.268327] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': True, 'boot_index': 0, 'mount_device': '/dev/sda', 'disk_bus': None, 'guest_format': None, 'device_type': None, 'attachment_id': '2d03c130-1c0c-4fe8-96b0-05438167c3d4', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994961', 'volume_id': 'eacd0d18-4608-43ee-a54c-6655db952ccb', 'name': 'volume-eacd0d18-4608-43ee-a54c-6655db952ccb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'daffdd40-9ee3-4a80-b670-5772395a32ae', 'attached_at': '', 'detached_at': '', 'volume_id': 'eacd0d18-4608-43ee-a54c-6655db952ccb', 'serial': 'eacd0d18-4608-43ee-a54c-6655db952ccb'}, 'volume_type': None}], 'swap': None} {{(pid=65522) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1210.268327] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Root volume attach. Driver type: vmdk {{(pid=65522) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1210.272056] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550c14ec-cb54-4610-ad0b-ee89435fde27 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.282736] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473559fa-3a0b-4b9a-a2dc-50ea72d01bc3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.294498] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1bfd2d2-55eb-48d7-a81e-f038686f1973 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.304099] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-585b4490-a1d4-485b-a762-7a946ed63502 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.323835] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1210.323835] env[65522]: value = "task-5115035" [ 1210.323835] env[65522]: _type = "Task" [ 1210.323835] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.331879] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115035, 'name': RelocateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.351410] env[65522]: DEBUG nova.network.neutron [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Updated VIF entry in instance network info cache for port 0c495486-a850-4368-95a9-03fac8060924. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1210.352256] env[65522]: DEBUG nova.network.neutron [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Updating instance_info_cache with network_info: [{"id": "0c495486-a850-4368-95a9-03fac8060924", "address": "fa:16:3e:a7:35:38", "network": {"id": "2eb73cd0-541f-42e9-8774-600c5c1e9200", "bridge": null, "label": "tempest-ServersNegativeTestJSON-939435024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdba2dbc660145b7a98af8bd3f77a035", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap0c495486-a8", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1210.354881] env[65522]: WARNING openstack [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1210.355267] env[65522]: WARNING openstack [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1210.375513] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115034, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.394927] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea32dbae-17dd-4b9a-9197-1ee4e01b42cc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.403851] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4c5858-d37a-4700-91c8-913e8e0313cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.437476] env[65522]: WARNING neutronclient.v2_0.client [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1210.438139] env[65522]: WARNING openstack [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1210.438511] env[65522]: WARNING openstack [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1210.447984] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064af501-0924-4430-b2be-91b305da18be {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.460010] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 75c8848b-aea9-43f1-8697-9224050d1fef] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1210.463482] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349bda98-9f9d-4fe0-ab80-9814cdb30316 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.480292] env[65522]: DEBUG nova.compute.provider_tree [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1210.542637] env[65522]: WARNING neutronclient.v2_0.client [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1210.569430] env[65522]: DEBUG nova.network.neutron [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updated VIF entry in instance network info cache for port 6b80135d-6228-4366-83aa-ff70808ae005. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1210.569837] env[65522]: DEBUG nova.network.neutron [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating instance_info_cache with network_info: [{"id": "6b80135d-6228-4366-83aa-ff70808ae005", "address": "fa:16:3e:82:79:a0", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b80135d-62", "ovs_interfaceid": "6b80135d-6228-4366-83aa-ff70808ae005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1210.602471] env[65522]: DEBUG nova.network.neutron [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Port 499e91eb-948a-4b54-b7a6-ce30af8315ea binding to destination host cpu-1 is already ACTIVE {{(pid=65522) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 1210.624902] env[65522]: WARNING openstack [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1210.625302] env[65522]: WARNING openstack [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1210.668653] env[65522]: DEBUG nova.network.neutron [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1210.690288] env[65522]: WARNING openstack [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1210.690754] env[65522]: WARNING openstack [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1210.765975] env[65522]: WARNING neutronclient.v2_0.client [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1210.766788] env[65522]: WARNING openstack [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1210.767179] env[65522]: WARNING openstack [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1210.834811] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115035, 'name': RelocateVM_Task} progress is 42%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.854867] env[65522]: DEBUG oslo_concurrency.lockutils [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Releasing lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1210.855054] env[65522]: DEBUG nova.compute.manager [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Received event network-vif-unplugged-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1210.855241] env[65522]: DEBUG oslo_concurrency.lockutils [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Acquiring lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1210.855507] env[65522]: DEBUG oslo_concurrency.lockutils [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1210.855676] env[65522]: DEBUG oslo_concurrency.lockutils [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1210.855919] env[65522]: DEBUG nova.compute.manager [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] No waiting events found dispatching network-vif-unplugged-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1210.856110] env[65522]: WARNING nova.compute.manager [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Received unexpected event network-vif-unplugged-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc for instance with vm_state shelved and task_state shelving_offloading. [ 1210.856275] env[65522]: DEBUG nova.compute.manager [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Received event network-changed-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1210.856429] env[65522]: DEBUG nova.compute.manager [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Refreshing instance network info cache due to event network-changed-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1210.856621] env[65522]: DEBUG oslo_concurrency.lockutils [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Acquiring lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.856780] env[65522]: DEBUG oslo_concurrency.lockutils [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Acquired lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1210.856941] env[65522]: DEBUG nova.network.neutron [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Refreshing network info cache for port fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1210.862325] env[65522]: DEBUG nova.network.neutron [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Updating instance_info_cache with network_info: [{"id": "0a9cf0d6-c5cb-4805-918d-ed092924fb2a", "address": "fa:16:3e:23:17:05", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a9cf0d6-c5", "ovs_interfaceid": "0a9cf0d6-c5cb-4805-918d-ed092924fb2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1210.879812] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115034, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.955560] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1210.963346] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: d5e8dd05-dc3c-4831-b4b0-ac100360f3e7] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1210.984047] env[65522]: DEBUG nova.scheduler.client.report [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1211.073466] env[65522]: DEBUG oslo_concurrency.lockutils [req-6c128ff0-2704-4863-be58-f65c65e4ed9e req-8659242c-e2e9-4d24-98b4-596641dd3450 service nova] Releasing lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1211.333805] env[65522]: DEBUG nova.compute.manager [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Received event network-vif-plugged-0a9cf0d6-c5cb-4805-918d-ed092924fb2a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1211.334536] env[65522]: DEBUG oslo_concurrency.lockutils [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] Acquiring lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1211.334796] env[65522]: DEBUG oslo_concurrency.lockutils [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] Lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.335035] env[65522]: DEBUG oslo_concurrency.lockutils [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] Lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.335177] env[65522]: DEBUG nova.compute.manager [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] No waiting events found dispatching network-vif-plugged-0a9cf0d6-c5cb-4805-918d-ed092924fb2a {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1211.335365] env[65522]: WARNING nova.compute.manager [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Received unexpected event network-vif-plugged-0a9cf0d6-c5cb-4805-918d-ed092924fb2a for instance with vm_state building and task_state spawning. [ 1211.335529] env[65522]: DEBUG nova.compute.manager [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Received event network-changed-0a9cf0d6-c5cb-4805-918d-ed092924fb2a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1211.335701] env[65522]: DEBUG nova.compute.manager [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Refreshing instance network info cache due to event network-changed-0a9cf0d6-c5cb-4805-918d-ed092924fb2a. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1211.335913] env[65522]: DEBUG oslo_concurrency.lockutils [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] Acquiring lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1211.341068] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115035, 'name': RelocateVM_Task} progress is 54%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.360894] env[65522]: WARNING neutronclient.v2_0.client [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1211.361770] env[65522]: WARNING openstack [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1211.362210] env[65522]: WARNING openstack [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1211.371814] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1211.372308] env[65522]: DEBUG nova.compute.manager [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Instance network_info: |[{"id": "0a9cf0d6-c5cb-4805-918d-ed092924fb2a", "address": "fa:16:3e:23:17:05", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a9cf0d6-c5", "ovs_interfaceid": "0a9cf0d6-c5cb-4805-918d-ed092924fb2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1211.372771] env[65522]: DEBUG oslo_concurrency.lockutils [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] Acquired lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1211.372952] env[65522]: DEBUG nova.network.neutron [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Refreshing network info cache for port 0a9cf0d6-c5cb-4805-918d-ed092924fb2a {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1211.374586] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:17:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4307c18-b235-43cd-bcd5-e226012d8ee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0a9cf0d6-c5cb-4805-918d-ed092924fb2a', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1211.383872] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1211.389085] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1211.389369] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a18a5494-cc02-43d5-a04b-efd045c0e6d6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.412724] env[65522]: DEBUG oslo_vmware.api [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115034, 'name': PowerOnVM_Task, 'duration_secs': 1.097472} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.413019] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1211.417313] env[65522]: DEBUG nova.compute.manager [None req-927f7636-6da3-413e-9f9d-5f4f7644adc3 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1211.418638] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b79b8a63-5193-4a93-98e2-95279e932e54 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.423307] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1211.423307] env[65522]: value = "task-5115036" [ 1211.423307] env[65522]: _type = "Task" [ 1211.423307] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.440038] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115036, 'name': CreateVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.468798] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 68d8d59b-4fc8-4993-8c9e-31b2329c88dd] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1211.489275] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.297s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.494048] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.350s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.494048] env[65522]: DEBUG nova.objects.instance [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lazy-loading 'resources' on Instance uuid 3b82f9ce-ea59-430d-adb6-918cc6cc48a4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1211.546029] env[65522]: WARNING openstack [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1211.546277] env[65522]: WARNING openstack [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1211.633494] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1211.633787] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.634887] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.646519] env[65522]: WARNING neutronclient.v2_0.client [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1211.647245] env[65522]: WARNING openstack [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1211.647568] env[65522]: WARNING openstack [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1211.752264] env[65522]: DEBUG nova.network.neutron [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Updated VIF entry in instance network info cache for port fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1211.753059] env[65522]: DEBUG nova.network.neutron [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Updating instance_info_cache with network_info: [{"id": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "address": "fa:16:3e:c3:dd:ad", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapfed73ed8-c0", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1211.838650] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115035, 'name': RelocateVM_Task} progress is 67%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.889828] env[65522]: WARNING neutronclient.v2_0.client [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1211.890629] env[65522]: WARNING openstack [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1211.891033] env[65522]: WARNING openstack [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1211.940176] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115036, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.973196] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 87a3c63d-794d-44ab-bad6-65c323d72ae7] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1211.996922] env[65522]: DEBUG nova.objects.instance [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lazy-loading 'numa_topology' on Instance uuid 3b82f9ce-ea59-430d-adb6-918cc6cc48a4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1212.001624] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fe59c6d-5206-46d3-9a90-28a821ceca1b tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 24.816s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1212.003045] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.048s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1212.003330] env[65522]: INFO nova.compute.manager [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Unshelving [ 1212.015861] env[65522]: WARNING openstack [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.016519] env[65522]: WARNING openstack [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.092135] env[65522]: WARNING neutronclient.v2_0.client [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1212.092135] env[65522]: WARNING openstack [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.092135] env[65522]: WARNING openstack [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.185450] env[65522]: DEBUG nova.network.neutron [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Updated VIF entry in instance network info cache for port 0a9cf0d6-c5cb-4805-918d-ed092924fb2a. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1212.185909] env[65522]: DEBUG nova.network.neutron [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Updating instance_info_cache with network_info: [{"id": "0a9cf0d6-c5cb-4805-918d-ed092924fb2a", "address": "fa:16:3e:23:17:05", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a9cf0d6-c5", "ovs_interfaceid": "0a9cf0d6-c5cb-4805-918d-ed092924fb2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1212.255566] env[65522]: DEBUG oslo_concurrency.lockutils [req-7137cc59-aefa-4e19-9798-e21f3b8e6683 req-85e68ead-3bc3-414d-bef5-c84c640f6ba2 service nova] Releasing lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1212.338796] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115035, 'name': RelocateVM_Task} progress is 81%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.436235] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115036, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.477397] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: ca964440-5375-4aff-8b45-96fbe829dd16] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1212.502489] env[65522]: DEBUG nova.objects.base [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Object Instance<3b82f9ce-ea59-430d-adb6-918cc6cc48a4> lazy-loaded attributes: resources,numa_topology {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1212.529921] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1212.640664] env[65522]: WARNING neutronclient.v2_0.client [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1212.672822] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e8857d-1fb6-4d02-9340-5a5f0eb15a29 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.676975] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.677204] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1212.677386] env[65522]: DEBUG nova.network.neutron [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1212.684534] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9888cdc-2125-4b03-b9e3-e5a5771981cd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.691833] env[65522]: DEBUG oslo_concurrency.lockutils [req-d2b047ac-7df5-40b5-8312-588c587401d9 req-2df582cf-ce89-4f26-8c8a-1f6e7cb2fb09 service nova] Releasing lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1212.724031] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9428b4ab-e463-4add-9eca-836fd12618af {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.733149] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195c3dcb-d402-46cb-b8a2-39db701b92c3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.749206] env[65522]: DEBUG nova.compute.provider_tree [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1212.837844] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115035, 'name': RelocateVM_Task} progress is 97%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.938778] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115036, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.981389] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: e5d2cc50-4ee6-4e64-9b52-888968a717ca] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1213.026837] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1213.181062] env[65522]: WARNING neutronclient.v2_0.client [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1213.181625] env[65522]: WARNING openstack [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1213.182094] env[65522]: WARNING openstack [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1213.252626] env[65522]: DEBUG nova.scheduler.client.report [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1213.282904] env[65522]: WARNING openstack [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1213.283338] env[65522]: WARNING openstack [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1213.340162] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115035, 'name': RelocateVM_Task} progress is 98%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.352914] env[65522]: WARNING neutronclient.v2_0.client [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1213.353804] env[65522]: WARNING openstack [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1213.354224] env[65522]: WARNING openstack [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1213.438735] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115036, 'name': CreateVM_Task, 'duration_secs': 1.537111} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.438959] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1213.439514] env[65522]: WARNING neutronclient.v2_0.client [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1213.439897] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.440066] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1213.440465] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1213.440655] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e08784d-a38b-4b7f-8890-a1160ce966d1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.444759] env[65522]: DEBUG nova.network.neutron [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating instance_info_cache with network_info: [{"id": "499e91eb-948a-4b54-b7a6-ce30af8315ea", "address": "fa:16:3e:09:07:e8", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap499e91eb-94", "ovs_interfaceid": "499e91eb-948a-4b54-b7a6-ce30af8315ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1213.450407] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1213.450407] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b86071-8ab0-a676-2805-2a8a9043a492" [ 1213.450407] env[65522]: _type = "Task" [ 1213.450407] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.463170] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b86071-8ab0-a676-2805-2a8a9043a492, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.484646] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 7c6f8218-602d-44f3-8012-de5a96972785] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1213.564916] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1213.565233] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1213.759368] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.266s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1213.762664] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.736s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1213.763056] env[65522]: DEBUG nova.objects.instance [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lazy-loading 'pci_requests' on Instance uuid 2981b635-43c8-4bd6-9991-e6af0be82f3c {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1213.839583] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115035, 'name': RelocateVM_Task, 'duration_secs': 3.232383} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.839851] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1213.840061] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994961', 'volume_id': 'eacd0d18-4608-43ee-a54c-6655db952ccb', 'name': 'volume-eacd0d18-4608-43ee-a54c-6655db952ccb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'daffdd40-9ee3-4a80-b670-5772395a32ae', 'attached_at': '', 'detached_at': '', 'volume_id': 'eacd0d18-4608-43ee-a54c-6655db952ccb', 'serial': 'eacd0d18-4608-43ee-a54c-6655db952ccb'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1213.841082] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c634a4-d528-4291-935b-a97f6da2512a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.857960] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e370e6a-d705-4752-b9ef-2cd544f561f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.885735] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] volume-eacd0d18-4608-43ee-a54c-6655db952ccb/volume-eacd0d18-4608-43ee-a54c-6655db952ccb.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1213.886297] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7fd75ad-a603-4e37-8b34-4cb403392890 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.916186] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1213.916186] env[65522]: value = "task-5115037" [ 1213.916186] env[65522]: _type = "Task" [ 1213.916186] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.924755] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115037, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.947613] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1213.961532] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b86071-8ab0-a676-2805-2a8a9043a492, 'name': SearchDatastore_Task, 'duration_secs': 0.02445} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.961843] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1213.962087] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1213.962321] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.962462] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1213.962639] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1213.962917] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f294c5ff-758d-4920-9ba4-75e5ad80104d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.972539] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1213.972822] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1213.974059] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-992e5a36-9623-4834-b00f-efa315363b27 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.980792] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1213.980792] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524aac35-27f5-35b2-3793-edc26d4fef1e" [ 1213.980792] env[65522]: _type = "Task" [ 1213.980792] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.990906] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524aac35-27f5-35b2-3793-edc26d4fef1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.067810] env[65522]: DEBUG nova.compute.manager [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1214.266062] env[65522]: DEBUG nova.objects.instance [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lazy-loading 'numa_topology' on Instance uuid 2981b635-43c8-4bd6-9991-e6af0be82f3c {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1214.268780] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ea87fbb6-7f3b-462c-af63-558ceff3723e tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.568s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1214.269659] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.740s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1214.269865] env[65522]: INFO nova.compute.manager [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Unshelving [ 1214.427187] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115037, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.475143] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab473b3-78f7-48d8-8306-727ed5cac18d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.497658] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dbe355a-98d4-4536-ae7c-d82585b4f609 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.508972] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating instance 'ae3455d8-2f38-42ad-b16d-d98aef92b2a8' progress to 83 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1214.512685] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]524aac35-27f5-35b2-3793-edc26d4fef1e, 'name': SearchDatastore_Task, 'duration_secs': 0.010113} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.513686] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63cb18de-dd97-4b5f-b697-bfaa9991b79f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.519478] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1214.519478] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529b76df-b691-d15a-f5e3-be90e067c174" [ 1214.519478] env[65522]: _type = "Task" [ 1214.519478] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.528405] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529b76df-b691-d15a-f5e3-be90e067c174, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.591503] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1214.769865] env[65522]: INFO nova.compute.claims [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1214.928731] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115037, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.014998] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1215.015868] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b2ebfbd-d636-415a-a445-c9973db692a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.025532] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1215.025532] env[65522]: value = "task-5115038" [ 1215.025532] env[65522]: _type = "Task" [ 1215.025532] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.033202] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529b76df-b691-d15a-f5e3-be90e067c174, 'name': SearchDatastore_Task, 'duration_secs': 0.012076} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.033939] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1215.034412] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 1b0fcdbb-aa9e-4915-a8c6-1167aa888390/1b0fcdbb-aa9e-4915-a8c6-1167aa888390.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1215.034826] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88203c90-5103-4bb6-a6bc-1bb190b1479b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.039647] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115038, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.046207] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1215.046207] env[65522]: value = "task-5115039" [ 1215.046207] env[65522]: _type = "Task" [ 1215.046207] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.056348] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115039, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.298885] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1215.434574] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115037, 'name': ReconfigVM_Task, 'duration_secs': 1.302162} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.434959] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Reconfigured VM instance instance-00000074 to attach disk [datastore1] volume-eacd0d18-4608-43ee-a54c-6655db952ccb/volume-eacd0d18-4608-43ee-a54c-6655db952ccb.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1215.440609] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea193071-e37e-4dde-bd8f-6372d13ceff8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.459638] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1215.459638] env[65522]: value = "task-5115040" [ 1215.459638] env[65522]: _type = "Task" [ 1215.459638] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.473515] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115040, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.537277] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115038, 'name': PowerOnVM_Task} progress is 90%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.559192] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115039, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.922767] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce8ae6e-d538-4160-9949-664f028483a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.931791] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a714d045-ea8c-453c-abf1-9eff179e6020 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.966027] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5184655-0d66-4622-a127-1524066aa624 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.977056] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc307a09-3335-43bd-b39a-8418a6500a62 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.980859] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115040, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.991249] env[65522]: DEBUG nova.compute.provider_tree [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1216.038372] env[65522]: DEBUG oslo_vmware.api [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115038, 'name': PowerOnVM_Task, 'duration_secs': 0.599964} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.038684] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1216.038852] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6ea625bf-b85d-4f80-8a9f-f03c5e1a4beb tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating instance 'ae3455d8-2f38-42ad-b16d-d98aef92b2a8' progress to 100 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1216.056039] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115039, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.54829} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.056039] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 1b0fcdbb-aa9e-4915-a8c6-1167aa888390/1b0fcdbb-aa9e-4915-a8c6-1167aa888390.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1216.056273] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1216.056324] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e29d5b19-cafe-4f3b-9b54-94f3696bb185 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.063714] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1216.063714] env[65522]: value = "task-5115041" [ 1216.063714] env[65522]: _type = "Task" [ 1216.063714] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.073905] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115041, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.472122] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115040, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.494311] env[65522]: DEBUG nova.scheduler.client.report [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1216.575048] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115041, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063681} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.575305] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1216.576134] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068207ad-5a57-413a-8452-6e568fcc566c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.599630] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] 1b0fcdbb-aa9e-4915-a8c6-1167aa888390/1b0fcdbb-aa9e-4915-a8c6-1167aa888390.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1216.600233] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9069957c-1e0d-43ac-a5c4-c5f53e7e1062 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.620152] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1216.620152] env[65522]: value = "task-5115042" [ 1216.620152] env[65522]: _type = "Task" [ 1216.620152] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.628995] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115042, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.974133] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115040, 'name': ReconfigVM_Task, 'duration_secs': 1.188615} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.974133] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994961', 'volume_id': 'eacd0d18-4608-43ee-a54c-6655db952ccb', 'name': 'volume-eacd0d18-4608-43ee-a54c-6655db952ccb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'daffdd40-9ee3-4a80-b670-5772395a32ae', 'attached_at': '', 'detached_at': '', 'volume_id': 'eacd0d18-4608-43ee-a54c-6655db952ccb', 'serial': 'eacd0d18-4608-43ee-a54c-6655db952ccb'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1216.974406] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bab2709a-600f-4e2c-a9ef-e3077a534590 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.981202] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1216.981202] env[65522]: value = "task-5115043" [ 1216.981202] env[65522]: _type = "Task" [ 1216.981202] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.990247] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115043, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.999206] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.236s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1216.999578] env[65522]: WARNING neutronclient.v2_0.client [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1217.002699] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.411s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1217.004024] env[65522]: INFO nova.compute.claims [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1217.093435] env[65522]: WARNING openstack [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1217.093872] env[65522]: WARNING openstack [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1217.177308] env[65522]: INFO nova.network.neutron [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Updating port 0c495486-a850-4368-95a9-03fac8060924 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1217.189688] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115042, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.491699] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115043, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.690141] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115042, 'name': ReconfigVM_Task, 'duration_secs': 0.76814} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.690519] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Reconfigured VM instance instance-00000075 to attach disk [datastore1] 1b0fcdbb-aa9e-4915-a8c6-1167aa888390/1b0fcdbb-aa9e-4915-a8c6-1167aa888390.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1217.691128] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f83f406a-d668-4dc5-ac5d-a2913844fc02 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.699184] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1217.699184] env[65522]: value = "task-5115044" [ 1217.699184] env[65522]: _type = "Task" [ 1217.699184] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.709405] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115044, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.906755] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1217.907060] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1217.907282] env[65522]: DEBUG nova.compute.manager [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Going to confirm migration 8 {{(pid=65522) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1217.994622] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115043, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.196189] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca89e926-e300-488c-9c96-6a22125aa611 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.211573] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abebe38a-5081-4d54-af41-2eab7f866c33 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.215257] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115044, 'name': Rename_Task, 'duration_secs': 0.160848} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.215520] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1218.216260] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e0d31a6-4267-4726-b2d0-6770a8452061 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.250095] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88d9e27-a557-4f09-bc1a-eb52e0fbd3be {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.253380] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1218.253380] env[65522]: value = "task-5115045" [ 1218.253380] env[65522]: _type = "Task" [ 1218.253380] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.260735] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3774d81e-4ad4-4800-b6b5-d8f41eb94ca7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.268806] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115045, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.280797] env[65522]: DEBUG nova.compute.provider_tree [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1218.413479] env[65522]: WARNING neutronclient.v2_0.client [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1218.451222] env[65522]: WARNING neutronclient.v2_0.client [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1218.451646] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.451829] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquired lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1218.452498] env[65522]: DEBUG nova.network.neutron [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1218.452498] env[65522]: DEBUG nova.objects.instance [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lazy-loading 'info_cache' on Instance uuid ae3455d8-2f38-42ad-b16d-d98aef92b2a8 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1218.494463] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115043, 'name': Rename_Task, 'duration_secs': 1.140132} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.494803] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1218.495133] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4f521663-3a91-4ccc-b5d2-9e13e24d9a1f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.503990] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1218.503990] env[65522]: value = "task-5115046" [ 1218.503990] env[65522]: _type = "Task" [ 1218.503990] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.515450] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115046, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.630722] env[65522]: DEBUG nova.compute.manager [req-89b1a6f0-3474-49cb-9c2a-89afb66100f3 req-8341f487-4f88-4561-9177-898cefee55b9 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Received event network-vif-plugged-0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1218.631049] env[65522]: DEBUG oslo_concurrency.lockutils [req-89b1a6f0-3474-49cb-9c2a-89afb66100f3 req-8341f487-4f88-4561-9177-898cefee55b9 service nova] Acquiring lock "2981b635-43c8-4bd6-9991-e6af0be82f3c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1218.631338] env[65522]: DEBUG oslo_concurrency.lockutils [req-89b1a6f0-3474-49cb-9c2a-89afb66100f3 req-8341f487-4f88-4561-9177-898cefee55b9 service nova] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1218.631581] env[65522]: DEBUG oslo_concurrency.lockutils [req-89b1a6f0-3474-49cb-9c2a-89afb66100f3 req-8341f487-4f88-4561-9177-898cefee55b9 service nova] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1218.631812] env[65522]: DEBUG nova.compute.manager [req-89b1a6f0-3474-49cb-9c2a-89afb66100f3 req-8341f487-4f88-4561-9177-898cefee55b9 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] No waiting events found dispatching network-vif-plugged-0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1218.632046] env[65522]: WARNING nova.compute.manager [req-89b1a6f0-3474-49cb-9c2a-89afb66100f3 req-8341f487-4f88-4561-9177-898cefee55b9 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Received unexpected event network-vif-plugged-0c495486-a850-4368-95a9-03fac8060924 for instance with vm_state shelved_offloaded and task_state spawning. [ 1218.709285] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.709880] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquired lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1218.709880] env[65522]: DEBUG nova.network.neutron [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1218.764719] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115045, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.784036] env[65522]: DEBUG nova.scheduler.client.report [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1219.016712] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115046, 'name': PowerOnVM_Task} progress is 74%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.213814] env[65522]: WARNING neutronclient.v2_0.client [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1219.214719] env[65522]: WARNING openstack [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1219.214995] env[65522]: WARNING openstack [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1219.265277] env[65522]: DEBUG oslo_vmware.api [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115045, 'name': PowerOnVM_Task, 'duration_secs': 0.716295} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.265592] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1219.265860] env[65522]: INFO nova.compute.manager [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Took 10.56 seconds to spawn the instance on the hypervisor. [ 1219.266094] env[65522]: DEBUG nova.compute.manager [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1219.267008] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff4d76c-5329-472e-9eb9-0a68cf2dabb8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.289081] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.286s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1219.289895] env[65522]: DEBUG nova.compute.manager [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1219.297805] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.999s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1219.298066] env[65522]: DEBUG nova.objects.instance [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lazy-loading 'pci_requests' on Instance uuid 3b82f9ce-ea59-430d-adb6-918cc6cc48a4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1219.331368] env[65522]: WARNING openstack [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1219.331812] env[65522]: WARNING openstack [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1219.404858] env[65522]: WARNING neutronclient.v2_0.client [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1219.405794] env[65522]: WARNING openstack [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1219.406181] env[65522]: WARNING openstack [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1219.459447] env[65522]: WARNING neutronclient.v2_0.client [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1219.460060] env[65522]: WARNING openstack [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1219.460633] env[65522]: WARNING openstack [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1219.515727] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115046, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.535139] env[65522]: DEBUG nova.network.neutron [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Updating instance_info_cache with network_info: [{"id": "0c495486-a850-4368-95a9-03fac8060924", "address": "fa:16:3e:a7:35:38", "network": {"id": "2eb73cd0-541f-42e9-8774-600c5c1e9200", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-939435024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdba2dbc660145b7a98af8bd3f77a035", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c495486-a8", "ovs_interfaceid": "0c495486-a850-4368-95a9-03fac8060924", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1219.570361] env[65522]: WARNING openstack [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1219.570865] env[65522]: WARNING openstack [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1219.656765] env[65522]: WARNING neutronclient.v2_0.client [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1219.657471] env[65522]: WARNING openstack [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1219.657819] env[65522]: WARNING openstack [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1219.762904] env[65522]: DEBUG nova.network.neutron [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating instance_info_cache with network_info: [{"id": "499e91eb-948a-4b54-b7a6-ce30af8315ea", "address": "fa:16:3e:09:07:e8", "network": {"id": "ccd9858f-4b40-49b4-8157-01d45d127d24", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1556392367-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f93c790ff61543bd8e134bcf9cb20bb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap499e91eb-94", "ovs_interfaceid": "499e91eb-948a-4b54-b7a6-ce30af8315ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1219.783493] env[65522]: INFO nova.compute.manager [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Took 15.63 seconds to build instance. [ 1219.801426] env[65522]: DEBUG nova.compute.utils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1219.803303] env[65522]: DEBUG nova.compute.manager [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1219.803525] env[65522]: DEBUG nova.network.neutron [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1219.803883] env[65522]: WARNING neutronclient.v2_0.client [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1219.804200] env[65522]: WARNING neutronclient.v2_0.client [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1219.804816] env[65522]: WARNING openstack [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1219.805175] env[65522]: WARNING openstack [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1219.815132] env[65522]: DEBUG nova.objects.instance [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lazy-loading 'numa_topology' on Instance uuid 3b82f9ce-ea59-430d-adb6-918cc6cc48a4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1219.851906] env[65522]: DEBUG nova.policy [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2f843db470614ed9b09fb5b6463827e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '15229af4da7a4ff9963369df50fb9d5b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1220.017629] env[65522]: DEBUG oslo_vmware.api [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115046, 'name': PowerOnVM_Task, 'duration_secs': 1.412379} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.019244] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1220.019244] env[65522]: INFO nova.compute.manager [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Took 11.40 seconds to spawn the instance on the hypervisor. [ 1220.019590] env[65522]: DEBUG nova.compute.manager [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1220.020650] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e551ad70-4353-4fb8-b991-4f9d5578ee17 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.040292] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Releasing lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1220.070584] env[65522]: DEBUG nova.virt.hardware [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='a5c8d1242f136e452e89d715275c6ea9',container_format='bare',created_at=2025-12-12T19:08:11Z,direct_url=,disk_format='vmdk',id=7832ad6e-429a-4894-9720-58fd3520c4e2,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-134801549-shelved',owner='fdba2dbc660145b7a98af8bd3f77a035',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2025-12-12T19:08:26Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1220.071233] env[65522]: DEBUG nova.virt.hardware [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1220.071466] env[65522]: DEBUG nova.virt.hardware [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1220.071781] env[65522]: DEBUG nova.virt.hardware [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1220.072056] env[65522]: DEBUG nova.virt.hardware [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1220.072364] env[65522]: DEBUG nova.virt.hardware [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1220.072620] env[65522]: DEBUG nova.virt.hardware [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1220.072923] env[65522]: DEBUG nova.virt.hardware [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1220.073285] env[65522]: DEBUG nova.virt.hardware [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1220.073492] env[65522]: DEBUG nova.virt.hardware [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1220.073769] env[65522]: DEBUG nova.virt.hardware [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1220.075136] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da1d720-6248-4f9b-9845-b86cea5e6a0a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.086105] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213b09a1-2902-4bea-830b-2fb6cefb7409 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.102779] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:35:38', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c495486-a850-4368-95a9-03fac8060924', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1220.111013] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1220.111329] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1220.111568] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c52a6e3-052b-48ae-a366-caf2553486db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.132450] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1220.132450] env[65522]: value = "task-5115047" [ 1220.132450] env[65522]: _type = "Task" [ 1220.132450] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.142864] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115047, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.143965] env[65522]: DEBUG nova.network.neutron [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Successfully created port: 6ff7afa0-2b0e-450c-a755-77a3241a2527 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1220.156747] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.156968] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.157128] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.157275] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.157419] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.157560] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.157698] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._sync_power_states {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.269038] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Releasing lock "refresh_cache-ae3455d8-2f38-42ad-b16d-d98aef92b2a8" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1220.269038] env[65522]: DEBUG nova.objects.instance [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lazy-loading 'migration_context' on Instance uuid ae3455d8-2f38-42ad-b16d-d98aef92b2a8 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1220.286931] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5c41e420-58f5-4e0d-b3c0-c8ba8c878b5b tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.140s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1220.314393] env[65522]: DEBUG nova.compute.manager [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1220.317992] env[65522]: INFO nova.compute.claims [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1220.541676] env[65522]: INFO nova.compute.manager [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Took 18.01 seconds to build instance. [ 1220.646328] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115047, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.662959] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Getting list of instances from cluster (obj){ [ 1220.662959] env[65522]: value = "domain-c8" [ 1220.662959] env[65522]: _type = "ClusterComputeResource" [ 1220.662959] env[65522]: } {{(pid=65522) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1220.664091] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4676a1ce-a3b6-4560-8c05-e942d9895868 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.682100] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Got total of 8 instances {{(pid=65522) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1220.682273] env[65522]: WARNING nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] While synchronizing instance power states, found 10 instances in the database and 8 instances on the hypervisor. [ 1220.682407] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Triggering sync for uuid 7617189c-a902-42e7-8165-0e7c4a1de06d {{(pid=65522) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1220.682711] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Triggering sync for uuid 2981b635-43c8-4bd6-9991-e6af0be82f3c {{(pid=65522) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1220.682941] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Triggering sync for uuid aebb7082-f88c-42f3-9d9f-93cb2a6f545b {{(pid=65522) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1220.683134] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Triggering sync for uuid 3987d71f-47ab-4dd4-8426-95d78503effa {{(pid=65522) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1220.683338] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Triggering sync for uuid ae3455d8-2f38-42ad-b16d-d98aef92b2a8 {{(pid=65522) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1220.683568] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Triggering sync for uuid 874b6a07-ebba-4f75-937e-2281af1598b8 {{(pid=65522) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1220.683779] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Triggering sync for uuid 3adfefa5-fb19-44b2-b3c5-42f2e2918673 {{(pid=65522) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1220.683981] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Triggering sync for uuid daffdd40-9ee3-4a80-b670-5772395a32ae {{(pid=65522) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1220.684211] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Triggering sync for uuid 1b0fcdbb-aa9e-4915-a8c6-1167aa888390 {{(pid=65522) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1220.684409] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Triggering sync for uuid 874813b4-d0d6-4559-82e2-9fa2ea4feb15 {{(pid=65522) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11044}} [ 1220.684858] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "7617189c-a902-42e7-8165-0e7c4a1de06d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.685092] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "7617189c-a902-42e7-8165-0e7c4a1de06d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.685379] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.685604] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.685781] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.686038] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "3987d71f-47ab-4dd4-8426-95d78503effa" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.686221] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.686463] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.686677] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "874b6a07-ebba-4f75-937e-2281af1598b8" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.686852] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "874b6a07-ebba-4f75-937e-2281af1598b8" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.687586] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.687809] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.688207] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "daffdd40-9ee3-4a80-b670-5772395a32ae" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.688457] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.688643] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1220.688889] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.689095] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.689227] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1220.689953] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676067d5-d35f-464c-9868-7b039005d1ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.693140] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e54afbe-9f2b-4f5c-bd2e-20d31ebcd75f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.695922] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f19442-632d-4623-a177-d591c5a5107b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.698648] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7898713f-9a21-4830-9a7e-d0f13f303d8d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.701378] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9f9712-9987-4c9e-8aa2-2d4ce638cd33 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.704057] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f59167f1-a83f-4fe6-991c-64ce94471062 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.707047] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.769273] env[65522]: DEBUG nova.objects.base [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1220.770384] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2efc4e5f-5456-4b52-8718-f7c04c68dafa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.792230] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c173ceb3-a6eb-44eb-a797-c2f2e60e6772 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.798810] env[65522]: DEBUG nova.compute.manager [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Received event network-changed-0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1220.799016] env[65522]: DEBUG nova.compute.manager [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Refreshing instance network info cache due to event network-changed-0c495486-a850-4368-95a9-03fac8060924. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1220.799237] env[65522]: DEBUG oslo_concurrency.lockutils [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Acquiring lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.799403] env[65522]: DEBUG oslo_concurrency.lockutils [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Acquired lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1220.799601] env[65522]: DEBUG nova.network.neutron [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Refreshing network info cache for port 0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1220.803741] env[65522]: DEBUG oslo_vmware.api [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1220.803741] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cac48d-9177-3e19-db55-fabe5335ffb1" [ 1220.803741] env[65522]: _type = "Task" [ 1220.803741] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.813639] env[65522]: DEBUG oslo_vmware.api [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cac48d-9177-3e19-db55-fabe5335ffb1, 'name': SearchDatastore_Task, 'duration_secs': 0.008465} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.813639] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1221.044624] env[65522]: DEBUG oslo_concurrency.lockutils [None req-30300263-b9e7-40fe-98ff-412c93f348e2 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "daffdd40-9ee3-4a80-b670-5772395a32ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.522s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1221.044962] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "daffdd40-9ee3-4a80-b670-5772395a32ae" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.357s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1221.045901] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e983b7f-1484-4695-a194-2bf1af84cd22 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.142981] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115047, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.210063] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1221.223945] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.536s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1221.226618] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.541s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1221.228128] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "7617189c-a902-42e7-8165-0e7c4a1de06d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.543s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1221.228487] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "874b6a07-ebba-4f75-937e-2281af1598b8" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.542s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1221.228871] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.543s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1221.235572] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.547s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1221.304773] env[65522]: WARNING neutronclient.v2_0.client [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1221.305511] env[65522]: WARNING openstack [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1221.305934] env[65522]: WARNING openstack [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1221.331536] env[65522]: DEBUG nova.compute.manager [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1221.366795] env[65522]: DEBUG nova.virt.hardware [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1221.367336] env[65522]: DEBUG nova.virt.hardware [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1221.367473] env[65522]: DEBUG nova.virt.hardware [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1221.367643] env[65522]: DEBUG nova.virt.hardware [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1221.367833] env[65522]: DEBUG nova.virt.hardware [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1221.368130] env[65522]: DEBUG nova.virt.hardware [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1221.368446] env[65522]: DEBUG nova.virt.hardware [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1221.368675] env[65522]: DEBUG nova.virt.hardware [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1221.371589] env[65522]: DEBUG nova.virt.hardware [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1221.371589] env[65522]: DEBUG nova.virt.hardware [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1221.371589] env[65522]: DEBUG nova.virt.hardware [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1221.372494] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a4a8a5-5115-4153-bde3-7601803e3027 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.384851] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d4d9bf-b6e8-4527-a579-49c0b708f894 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.536241] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f4cf15e-45f6-48fb-b11b-6c6760573222 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.545597] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b34c33-264f-4f28-b55a-02fe3ddf9721 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.578872] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "daffdd40-9ee3-4a80-b670-5772395a32ae" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.534s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1221.583068] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00252fac-1790-49fb-abc6-a72e771fc745 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.591705] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdeba662-c693-4ae8-996f-9eef7c86d062 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.606627] env[65522]: DEBUG nova.compute.provider_tree [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1221.648641] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115047, 'name': CreateVM_Task, 'duration_secs': 1.331935} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.649174] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1221.649697] env[65522]: WARNING neutronclient.v2_0.client [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1221.650370] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7832ad6e-429a-4894-9720-58fd3520c4e2" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.650370] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7832ad6e-429a-4894-9720-58fd3520c4e2" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1221.650644] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7832ad6e-429a-4894-9720-58fd3520c4e2" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1221.650968] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bf1bc6b-7d2d-4129-8429-26b020b91db0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.656517] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1221.656517] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c0ac23-98e3-7630-f0ff-7349af4b2d6b" [ 1221.656517] env[65522]: _type = "Task" [ 1221.656517] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.666023] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c0ac23-98e3-7630-f0ff-7349af4b2d6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.677325] env[65522]: WARNING openstack [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1221.677737] env[65522]: WARNING openstack [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1221.728453] env[65522]: DEBUG nova.network.neutron [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Successfully updated port: 6ff7afa0-2b0e-450c-a755-77a3241a2527 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1221.761990] env[65522]: WARNING neutronclient.v2_0.client [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1221.762803] env[65522]: WARNING openstack [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1221.763281] env[65522]: WARNING openstack [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1221.862621] env[65522]: DEBUG nova.network.neutron [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Updated VIF entry in instance network info cache for port 0c495486-a850-4368-95a9-03fac8060924. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1221.863107] env[65522]: DEBUG nova.network.neutron [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Updating instance_info_cache with network_info: [{"id": "0c495486-a850-4368-95a9-03fac8060924", "address": "fa:16:3e:a7:35:38", "network": {"id": "2eb73cd0-541f-42e9-8774-600c5c1e9200", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-939435024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdba2dbc660145b7a98af8bd3f77a035", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c495486-a8", "ovs_interfaceid": "0c495486-a850-4368-95a9-03fac8060924", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1221.909385] env[65522]: DEBUG nova.compute.manager [req-298c4213-241b-418c-a8f4-15a94522b250 req-f7933ecf-84b7-4630-9ae2-96ded2d5a170 service nova] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Received event network-vif-plugged-6ff7afa0-2b0e-450c-a755-77a3241a2527 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1221.909610] env[65522]: DEBUG oslo_concurrency.lockutils [req-298c4213-241b-418c-a8f4-15a94522b250 req-f7933ecf-84b7-4630-9ae2-96ded2d5a170 service nova] Acquiring lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1221.910048] env[65522]: DEBUG oslo_concurrency.lockutils [req-298c4213-241b-418c-a8f4-15a94522b250 req-f7933ecf-84b7-4630-9ae2-96ded2d5a170 service nova] Lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1221.910290] env[65522]: DEBUG oslo_concurrency.lockutils [req-298c4213-241b-418c-a8f4-15a94522b250 req-f7933ecf-84b7-4630-9ae2-96ded2d5a170 service nova] Lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1221.910493] env[65522]: DEBUG nova.compute.manager [req-298c4213-241b-418c-a8f4-15a94522b250 req-f7933ecf-84b7-4630-9ae2-96ded2d5a170 service nova] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] No waiting events found dispatching network-vif-plugged-6ff7afa0-2b0e-450c-a755-77a3241a2527 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1221.910676] env[65522]: WARNING nova.compute.manager [req-298c4213-241b-418c-a8f4-15a94522b250 req-f7933ecf-84b7-4630-9ae2-96ded2d5a170 service nova] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Received unexpected event network-vif-plugged-6ff7afa0-2b0e-450c-a755-77a3241a2527 for instance with vm_state building and task_state spawning. [ 1222.109632] env[65522]: DEBUG nova.scheduler.client.report [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1222.168532] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7832ad6e-429a-4894-9720-58fd3520c4e2" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1222.168801] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Processing image 7832ad6e-429a-4894-9720-58fd3520c4e2 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1222.169063] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7832ad6e-429a-4894-9720-58fd3520c4e2/7832ad6e-429a-4894-9720-58fd3520c4e2.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.169213] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7832ad6e-429a-4894-9720-58fd3520c4e2/7832ad6e-429a-4894-9720-58fd3520c4e2.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1222.169396] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1222.169978] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7d3e5b32-3c9e-44be-8cf2-797877656915 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.192501] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1222.192706] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1222.193638] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b4f6649-d52b-4514-a954-6f370efb2c06 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.199910] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1222.199910] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527a8cff-ccdb-8a61-4e01-dfcd3a39a612" [ 1222.199910] env[65522]: _type = "Task" [ 1222.199910] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.211646] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527a8cff-ccdb-8a61-4e01-dfcd3a39a612, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.234532] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "refresh_cache-874813b4-d0d6-4559-82e2-9fa2ea4feb15" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.234727] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquired lock "refresh_cache-874813b4-d0d6-4559-82e2-9fa2ea4feb15" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1222.234909] env[65522]: DEBUG nova.network.neutron [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1222.366390] env[65522]: DEBUG oslo_concurrency.lockutils [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Releasing lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1222.366724] env[65522]: DEBUG nova.compute.manager [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Received event network-changed-0a9cf0d6-c5cb-4805-918d-ed092924fb2a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1222.366840] env[65522]: DEBUG nova.compute.manager [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Refreshing instance network info cache due to event network-changed-0a9cf0d6-c5cb-4805-918d-ed092924fb2a. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1222.367072] env[65522]: DEBUG oslo_concurrency.lockutils [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Acquiring lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.367216] env[65522]: DEBUG oslo_concurrency.lockutils [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Acquired lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1222.367428] env[65522]: DEBUG nova.network.neutron [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Refreshing network info cache for port 0a9cf0d6-c5cb-4805-918d-ed092924fb2a {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1222.615764] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.318s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1222.616203] env[65522]: WARNING neutronclient.v2_0.client [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1222.619284] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.805s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1222.648874] env[65522]: INFO nova.network.neutron [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Updating port fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1222.714234] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Preparing fetch location {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1222.714234] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Fetch image to [datastore2] OSTACK_IMG_cab48bd4-4166-4507-af63-a2c2a1d58384/OSTACK_IMG_cab48bd4-4166-4507-af63-a2c2a1d58384.vmdk {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1222.714234] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Downloading stream optimized image 7832ad6e-429a-4894-9720-58fd3520c4e2 to [datastore2] OSTACK_IMG_cab48bd4-4166-4507-af63-a2c2a1d58384/OSTACK_IMG_cab48bd4-4166-4507-af63-a2c2a1d58384.vmdk on the data store datastore2 as vApp {{(pid=65522) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1222.714234] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Downloading image file data 7832ad6e-429a-4894-9720-58fd3520c4e2 to the ESX as VM named 'OSTACK_IMG_cab48bd4-4166-4507-af63-a2c2a1d58384' {{(pid=65522) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1222.738416] env[65522]: WARNING openstack [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1222.738836] env[65522]: WARNING openstack [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1222.788278] env[65522]: DEBUG nova.network.neutron [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1222.804755] env[65522]: DEBUG oslo_vmware.rw_handles [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1222.804755] env[65522]: value = "resgroup-9" [ 1222.804755] env[65522]: _type = "ResourcePool" [ 1222.804755] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1222.805525] env[65522]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-8b7da8fe-f32b-45cb-aff6-b1bce027bf1a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.842483] env[65522]: WARNING openstack [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1222.843047] env[65522]: WARNING openstack [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1222.855608] env[65522]: DEBUG nova.compute.manager [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Received event network-changed-dc729a30-fef3-4b2e-ab41-6c7a4eb89f73 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1222.855860] env[65522]: DEBUG nova.compute.manager [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Refreshing instance network info cache due to event network-changed-dc729a30-fef3-4b2e-ab41-6c7a4eb89f73. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1222.856097] env[65522]: DEBUG oslo_concurrency.lockutils [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Acquiring lock "refresh_cache-7617189c-a902-42e7-8165-0e7c4a1de06d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.856240] env[65522]: DEBUG oslo_concurrency.lockutils [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Acquired lock "refresh_cache-7617189c-a902-42e7-8165-0e7c4a1de06d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1222.856392] env[65522]: DEBUG nova.network.neutron [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Refreshing network info cache for port dc729a30-fef3-4b2e-ab41-6c7a4eb89f73 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1222.859570] env[65522]: DEBUG oslo_vmware.rw_handles [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lease: (returnval){ [ 1222.859570] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a9ab89-d4e8-3472-0433-4c27d7f21973" [ 1222.859570] env[65522]: _type = "HttpNfcLease" [ 1222.859570] env[65522]: } obtained for vApp import into resource pool (val){ [ 1222.859570] env[65522]: value = "resgroup-9" [ 1222.859570] env[65522]: _type = "ResourcePool" [ 1222.859570] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1222.859993] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the lease: (returnval){ [ 1222.859993] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a9ab89-d4e8-3472-0433-4c27d7f21973" [ 1222.859993] env[65522]: _type = "HttpNfcLease" [ 1222.859993] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1222.868856] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1222.868856] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a9ab89-d4e8-3472-0433-4c27d7f21973" [ 1222.868856] env[65522]: _type = "HttpNfcLease" [ 1222.868856] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1222.870510] env[65522]: WARNING neutronclient.v2_0.client [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1222.871155] env[65522]: WARNING openstack [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1222.871511] env[65522]: WARNING openstack [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1222.924046] env[65522]: WARNING neutronclient.v2_0.client [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1222.924819] env[65522]: WARNING openstack [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1222.925163] env[65522]: WARNING openstack [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.085720] env[65522]: WARNING openstack [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.086185] env[65522]: WARNING openstack [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.117681] env[65522]: DEBUG nova.network.neutron [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Updating instance_info_cache with network_info: [{"id": "6ff7afa0-2b0e-450c-a755-77a3241a2527", "address": "fa:16:3e:c7:7a:9b", "network": {"id": "26dda233-0dd0-40ae-b56f-c84b1b91e92a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1197499239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "15229af4da7a4ff9963369df50fb9d5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ff7afa0-2b", "ovs_interfaceid": "6ff7afa0-2b0e-450c-a755-77a3241a2527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1223.250746] env[65522]: WARNING neutronclient.v2_0.client [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1223.251403] env[65522]: WARNING openstack [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.251754] env[65522]: WARNING openstack [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.309808] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-332d9479-5e16-4f8d-9eb6-131579bd3507 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.322379] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fd951e-f0f7-4102-8dea-f8c41d551120 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.361729] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c136402-4b25-424e-8587-67b32e3bfacf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.364913] env[65522]: WARNING neutronclient.v2_0.client [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1223.365760] env[65522]: WARNING openstack [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.366078] env[65522]: WARNING openstack [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.375120] env[65522]: DEBUG nova.network.neutron [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Updated VIF entry in instance network info cache for port 0a9cf0d6-c5cb-4805-918d-ed092924fb2a. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1223.375714] env[65522]: DEBUG nova.network.neutron [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Updating instance_info_cache with network_info: [{"id": "0a9cf0d6-c5cb-4805-918d-ed092924fb2a", "address": "fa:16:3e:23:17:05", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a9cf0d6-c5", "ovs_interfaceid": "0a9cf0d6-c5cb-4805-918d-ed092924fb2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1223.379795] env[65522]: DEBUG nova.compute.manager [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Stashing vm_state: active {{(pid=65522) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1223.391885] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feccd47c-cf34-4e0f-b88f-984dc998877d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.395915] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1223.395915] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a9ab89-d4e8-3472-0433-4c27d7f21973" [ 1223.395915] env[65522]: _type = "HttpNfcLease" [ 1223.395915] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1223.408488] env[65522]: DEBUG nova.compute.provider_tree [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1223.547391] env[65522]: WARNING openstack [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.547911] env[65522]: WARNING openstack [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.620293] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Releasing lock "refresh_cache-874813b4-d0d6-4559-82e2-9fa2ea4feb15" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1223.620649] env[65522]: DEBUG nova.compute.manager [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Instance network_info: |[{"id": "6ff7afa0-2b0e-450c-a755-77a3241a2527", "address": "fa:16:3e:c7:7a:9b", "network": {"id": "26dda233-0dd0-40ae-b56f-c84b1b91e92a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1197499239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "15229af4da7a4ff9963369df50fb9d5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ff7afa0-2b", "ovs_interfaceid": "6ff7afa0-2b0e-450c-a755-77a3241a2527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1223.621147] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c7:7a:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa8c2f93-f287-41b3-adb6-4942a7ea2a0b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ff7afa0-2b0e-450c-a755-77a3241a2527', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1223.629074] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1223.630417] env[65522]: WARNING neutronclient.v2_0.client [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1223.631080] env[65522]: WARNING openstack [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.631444] env[65522]: WARNING openstack [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.639824] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1223.640410] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ecdc0c2-3af1-4deb-9e03-031f3534b48a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.661305] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1223.661305] env[65522]: value = "task-5115049" [ 1223.661305] env[65522]: _type = "Task" [ 1223.661305] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.671306] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115049, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.722403] env[65522]: DEBUG nova.network.neutron [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Updated VIF entry in instance network info cache for port dc729a30-fef3-4b2e-ab41-6c7a4eb89f73. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1223.722775] env[65522]: DEBUG nova.network.neutron [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Updating instance_info_cache with network_info: [{"id": "dc729a30-fef3-4b2e-ab41-6c7a4eb89f73", "address": "fa:16:3e:6a:3b:a6", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc729a30-fe", "ovs_interfaceid": "dc729a30-fef3-4b2e-ab41-6c7a4eb89f73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1223.882169] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1223.882169] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a9ab89-d4e8-3472-0433-4c27d7f21973" [ 1223.882169] env[65522]: _type = "HttpNfcLease" [ 1223.882169] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1223.884906] env[65522]: DEBUG oslo_concurrency.lockutils [req-a3ec3509-9a6f-4dc1-9868-86a92c8c1e29 req-bda7139c-591e-4718-9a84-28b1b5ee8803 service nova] Releasing lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1223.903500] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1223.911826] env[65522]: DEBUG nova.scheduler.client.report [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1223.982588] env[65522]: DEBUG nova.compute.manager [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Received event network-changed-6ff7afa0-2b0e-450c-a755-77a3241a2527 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1223.982790] env[65522]: DEBUG nova.compute.manager [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Refreshing instance network info cache due to event network-changed-6ff7afa0-2b0e-450c-a755-77a3241a2527. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1223.983031] env[65522]: DEBUG oslo_concurrency.lockutils [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] Acquiring lock "refresh_cache-874813b4-d0d6-4559-82e2-9fa2ea4feb15" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1223.983178] env[65522]: DEBUG oslo_concurrency.lockutils [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] Acquired lock "refresh_cache-874813b4-d0d6-4559-82e2-9fa2ea4feb15" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1223.983339] env[65522]: DEBUG nova.network.neutron [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Refreshing network info cache for port 6ff7afa0-2b0e-450c-a755-77a3241a2527 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1224.172369] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115049, 'name': CreateVM_Task, 'duration_secs': 0.360586} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.172512] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1224.173038] env[65522]: WARNING neutronclient.v2_0.client [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1224.173741] env[65522]: WARNING openstack [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.174113] env[65522]: WARNING openstack [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.249642] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1224.249833] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1224.250193] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1224.250731] env[65522]: DEBUG oslo_concurrency.lockutils [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Releasing lock "refresh_cache-7617189c-a902-42e7-8165-0e7c4a1de06d" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1224.250958] env[65522]: DEBUG nova.compute.manager [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Received event network-changed-6b80135d-6228-4366-83aa-ff70808ae005 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1224.251137] env[65522]: DEBUG nova.compute.manager [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Refreshing instance network info cache due to event network-changed-6b80135d-6228-4366-83aa-ff70808ae005. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1224.251328] env[65522]: DEBUG oslo_concurrency.lockutils [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Acquiring lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1224.251469] env[65522]: DEBUG oslo_concurrency.lockutils [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Acquired lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1224.251687] env[65522]: DEBUG nova.network.neutron [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Refreshing network info cache for port 6b80135d-6228-4366-83aa-ff70808ae005 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1224.253134] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-008c00da-a114-4108-972d-b03a2b5434dd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.255784] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1224.255949] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1224.256121] env[65522]: DEBUG nova.network.neutron [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1224.263876] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1224.263876] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cbb958-570f-dca1-8361-6f54e9f5f296" [ 1224.263876] env[65522]: _type = "Task" [ 1224.263876] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.273410] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cbb958-570f-dca1-8361-6f54e9f5f296, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.383162] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1224.383162] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a9ab89-d4e8-3472-0433-4c27d7f21973" [ 1224.383162] env[65522]: _type = "HttpNfcLease" [ 1224.383162] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1224.383603] env[65522]: DEBUG oslo_vmware.rw_handles [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1224.383603] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a9ab89-d4e8-3472-0433-4c27d7f21973" [ 1224.383603] env[65522]: _type = "HttpNfcLease" [ 1224.383603] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1224.384377] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6021e24-0f0b-4570-bb24-f61163af1147 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.391823] env[65522]: DEBUG oslo_vmware.rw_handles [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525eed9c-dfb8-37e0-73ea-edf3d2a7f8e7/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1224.392035] env[65522]: DEBUG oslo_vmware.rw_handles [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525eed9c-dfb8-37e0-73ea-edf3d2a7f8e7/disk-0.vmdk. {{(pid=65522) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1224.454282] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e6d21903-8c99-4e60-b468-debff8ea6287 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.486811] env[65522]: WARNING neutronclient.v2_0.client [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1224.487502] env[65522]: WARNING openstack [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.487922] env[65522]: WARNING openstack [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.598808] env[65522]: WARNING openstack [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.599289] env[65522]: WARNING openstack [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.650846] env[65522]: WARNING neutronclient.v2_0.client [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1224.651762] env[65522]: WARNING openstack [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.652210] env[65522]: WARNING openstack [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.746432] env[65522]: DEBUG nova.network.neutron [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Updated VIF entry in instance network info cache for port 6ff7afa0-2b0e-450c-a755-77a3241a2527. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1224.746798] env[65522]: DEBUG nova.network.neutron [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Updating instance_info_cache with network_info: [{"id": "6ff7afa0-2b0e-450c-a755-77a3241a2527", "address": "fa:16:3e:c7:7a:9b", "network": {"id": "26dda233-0dd0-40ae-b56f-c84b1b91e92a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1197499239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "15229af4da7a4ff9963369df50fb9d5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ff7afa0-2b", "ovs_interfaceid": "6ff7afa0-2b0e-450c-a755-77a3241a2527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1224.759944] env[65522]: WARNING neutronclient.v2_0.client [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1224.760730] env[65522]: WARNING openstack [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.761088] env[65522]: WARNING openstack [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.773024] env[65522]: WARNING neutronclient.v2_0.client [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1224.773024] env[65522]: WARNING openstack [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.773024] env[65522]: WARNING openstack [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.793452] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cbb958-570f-dca1-8361-6f54e9f5f296, 'name': SearchDatastore_Task, 'duration_secs': 0.015988} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.795814] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1224.796568] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1224.796568] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1224.796568] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1224.796844] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1224.797330] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8a2307eb-2006-42ed-b19a-1c8667472660 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.808923] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1224.809155] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1224.813823] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a8ab2a1-2dca-430c-9fff-56fc8977e54a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.825781] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1224.825781] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f325b2-c5dc-1f40-bf7b-2e4141d75ffa" [ 1224.825781] env[65522]: _type = "Task" [ 1224.825781] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.836963] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f325b2-c5dc-1f40-bf7b-2e4141d75ffa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.887774] env[65522]: DEBUG nova.compute.manager [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Received event network-vif-plugged-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1224.888061] env[65522]: DEBUG oslo_concurrency.lockutils [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] Acquiring lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1224.888271] env[65522]: DEBUG oslo_concurrency.lockutils [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1224.888441] env[65522]: DEBUG oslo_concurrency.lockutils [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1224.888609] env[65522]: DEBUG nova.compute.manager [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] No waiting events found dispatching network-vif-plugged-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1224.888810] env[65522]: WARNING nova.compute.manager [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Received unexpected event network-vif-plugged-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc for instance with vm_state shelved_offloaded and task_state spawning. [ 1224.889018] env[65522]: DEBUG nova.compute.manager [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Received event network-changed-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1224.889186] env[65522]: DEBUG nova.compute.manager [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Refreshing instance network info cache due to event network-changed-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1224.889352] env[65522]: DEBUG oslo_concurrency.lockutils [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] Acquiring lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1224.905503] env[65522]: WARNING openstack [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.905880] env[65522]: WARNING openstack [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.918583] env[65522]: WARNING openstack [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.918956] env[65522]: WARNING openstack [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.956190] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.337s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1224.959226] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.749s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1224.960190] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1224.960190] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1224.960190] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.057s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1224.964061] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd99563c-96dc-4860-91bf-0f8a9523592e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.971983] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ba0220-53a0-4462-99f4-fe78d997f281 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.980573] env[65522]: WARNING neutronclient.v2_0.client [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1224.981985] env[65522]: WARNING openstack [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.981985] env[65522]: WARNING openstack [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1225.002644] env[65522]: WARNING neutronclient.v2_0.client [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1225.003403] env[65522]: WARNING openstack [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1225.003833] env[65522]: WARNING openstack [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1225.013390] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-792e1586-42c4-4c6a-bfcd-c4934a5012ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.029655] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf03949-a01a-4b33-9342-eaa7327c2aaa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.072333] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179613MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1225.072459] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.085619] env[65522]: DEBUG nova.network.neutron [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Updating instance_info_cache with network_info: [{"id": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "address": "fa:16:3e:c3:dd:ad", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfed73ed8-c0", "ovs_interfaceid": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1225.127751] env[65522]: DEBUG nova.network.neutron [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updated VIF entry in instance network info cache for port 6b80135d-6228-4366-83aa-ff70808ae005. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1225.128161] env[65522]: DEBUG nova.network.neutron [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating instance_info_cache with network_info: [{"id": "6b80135d-6228-4366-83aa-ff70808ae005", "address": "fa:16:3e:82:79:a0", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b80135d-62", "ovs_interfaceid": "6b80135d-6228-4366-83aa-ff70808ae005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1225.250417] env[65522]: DEBUG oslo_concurrency.lockutils [req-5721774b-a3f1-4eaa-922f-258a9f2d637a req-c2baf928-8c74-4be1-9892-dfe7bce361bd service nova] Releasing lock "refresh_cache-874813b4-d0d6-4559-82e2-9fa2ea4feb15" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1225.338046] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f325b2-c5dc-1f40-bf7b-2e4141d75ffa, 'name': SearchDatastore_Task, 'duration_secs': 0.010495} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.338905] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af10da3f-9a7e-4dba-bd6a-764cfb434a80 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.347795] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1225.347795] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bf0a84-623a-5ca9-c011-5db0860ac4a8" [ 1225.347795] env[65522]: _type = "Task" [ 1225.347795] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.359424] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bf0a84-623a-5ca9-c011-5db0860ac4a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.470567] env[65522]: INFO nova.compute.claims [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1225.536323] env[65522]: INFO nova.scheduler.client.report [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleted allocation for migration afd6f5d0-9391-4e8c-9e6f-16d7216ef2c5 [ 1225.589756] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1225.593251] env[65522]: DEBUG oslo_concurrency.lockutils [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] Acquired lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1225.593521] env[65522]: DEBUG nova.network.neutron [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Refreshing network info cache for port fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1225.618500] env[65522]: DEBUG nova.virt.hardware [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='68882ea6675184006603b2d6af0c3b44',container_format='bare',created_at=2025-12-12T19:08:11Z,direct_url=,disk_format='vmdk',id=e4e8fd5f-0bc1-437d-a8db-457c5ea05344,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-521064632-shelved',owner='1546bbdca07c443d84abea1155cfde37',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2025-12-12T19:08:28Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1225.618744] env[65522]: DEBUG nova.virt.hardware [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1225.618899] env[65522]: DEBUG nova.virt.hardware [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1225.619133] env[65522]: DEBUG nova.virt.hardware [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1225.619390] env[65522]: DEBUG nova.virt.hardware [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1225.619591] env[65522]: DEBUG nova.virt.hardware [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1225.619843] env[65522]: DEBUG nova.virt.hardware [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1225.620007] env[65522]: DEBUG nova.virt.hardware [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1225.620207] env[65522]: DEBUG nova.virt.hardware [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1225.620374] env[65522]: DEBUG nova.virt.hardware [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1225.620550] env[65522]: DEBUG nova.virt.hardware [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1225.621870] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32b9a0b-821d-42f3-8268-ac52972a310d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.634900] env[65522]: DEBUG oslo_concurrency.lockutils [req-5586d2f8-1b5b-439b-babf-82a987099859 req-b4bb74a3-5878-4273-98ce-36fb671e63d6 service nova] Releasing lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1225.638010] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d055feda-4a43-4a85-99d2-534f85d2a287 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.659430] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:dd:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4c5eb94-841c-4713-985a-8fc4117fbaf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1225.671096] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1225.675322] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1225.676272] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-79e1347d-b622-4984-8e76-884ff85ed127 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.698759] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1225.698759] env[65522]: value = "task-5115050" [ 1225.698759] env[65522]: _type = "Task" [ 1225.698759] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.707133] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115050, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.760740] env[65522]: DEBUG oslo_vmware.rw_handles [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Completed reading data from the image iterator. {{(pid=65522) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1225.760971] env[65522]: DEBUG oslo_vmware.rw_handles [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525eed9c-dfb8-37e0-73ea-edf3d2a7f8e7/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1225.762262] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4133d3-6aab-4d17-8b11-a8959a185f62 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.770506] env[65522]: DEBUG oslo_vmware.rw_handles [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525eed9c-dfb8-37e0-73ea-edf3d2a7f8e7/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1225.770730] env[65522]: DEBUG oslo_vmware.rw_handles [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525eed9c-dfb8-37e0-73ea-edf3d2a7f8e7/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1225.771013] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ce5e7fc6-c878-4ce4-9deb-aafaaa6eccdc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.859188] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bf0a84-623a-5ca9-c011-5db0860ac4a8, 'name': SearchDatastore_Task, 'duration_secs': 0.011065} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.859478] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1225.859774] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 874813b4-d0d6-4559-82e2-9fa2ea4feb15/874813b4-d0d6-4559-82e2-9fa2ea4feb15.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1225.860080] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d5c2cd07-22e1-43bf-8922-dbfdec6afc49 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.868599] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1225.868599] env[65522]: value = "task-5115051" [ 1225.868599] env[65522]: _type = "Task" [ 1225.868599] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.877680] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115051, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.980407] env[65522]: INFO nova.compute.resource_tracker [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating resource usage from migration 5817abdf-1156-4dda-8607-b249c325820b [ 1226.043540] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.136s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.045104] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.358s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1226.045377] env[65522]: INFO nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] During sync_power_state the instance has a pending task (deleting). Skip. [ 1226.045577] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.051393] env[65522]: DEBUG oslo_vmware.rw_handles [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525eed9c-dfb8-37e0-73ea-edf3d2a7f8e7/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1226.051684] env[65522]: INFO nova.virt.vmwareapi.images [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Downloaded image file data 7832ad6e-429a-4894-9720-58fd3520c4e2 [ 1226.052708] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-770dbb31-fabd-4b71-a9f3-8026b3d48258 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.073049] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-834f40d3-9f07-45c9-b622-8356fb8f6632 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.096974] env[65522]: WARNING neutronclient.v2_0.client [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1226.097653] env[65522]: WARNING openstack [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1226.098047] env[65522]: WARNING openstack [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1226.109188] env[65522]: INFO nova.virt.vmwareapi.images [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] The imported VM was unregistered [ 1226.112013] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Caching image {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1226.112278] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Creating directory with path [datastore2] devstack-image-cache_base/7832ad6e-429a-4894-9720-58fd3520c4e2 {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1226.113118] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-529ff991-91fa-4ed6-ac2e-1f0014d9d80a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.157592] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Created directory with path [datastore2] devstack-image-cache_base/7832ad6e-429a-4894-9720-58fd3520c4e2 {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1226.157882] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_cab48bd4-4166-4507-af63-a2c2a1d58384/OSTACK_IMG_cab48bd4-4166-4507-af63-a2c2a1d58384.vmdk to [datastore2] devstack-image-cache_base/7832ad6e-429a-4894-9720-58fd3520c4e2/7832ad6e-429a-4894-9720-58fd3520c4e2.vmdk. {{(pid=65522) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1226.158219] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-e0710f33-3fe8-4729-aacf-7e4f398f76d4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.171582] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1226.171582] env[65522]: value = "task-5115053" [ 1226.171582] env[65522]: _type = "Task" [ 1226.171582] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.186139] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115053, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.197434] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecbc9697-09a1-4cd4-bed5-0604cee80014 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.212473] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115050, 'name': CreateVM_Task, 'duration_secs': 0.355254} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.214820] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1226.215581] env[65522]: WARNING neutronclient.v2_0.client [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1226.215995] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e4e8fd5f-0bc1-437d-a8db-457c5ea05344" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1226.216179] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e4e8fd5f-0bc1-437d-a8db-457c5ea05344" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1226.216654] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e4e8fd5f-0bc1-437d-a8db-457c5ea05344" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1226.217816] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39cb0848-c965-4197-8b7e-d4a6d9ca1f91 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.222030] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d8282c2-d6ec-48d4-9978-9ce4f155b27c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.264215] env[65522]: WARNING openstack [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1226.264707] env[65522]: WARNING openstack [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1226.272583] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc0c01f-2e4e-4c4c-84a1-ccdfbaa0dc5f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.276122] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1226.276122] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52df21ff-5b7c-ac09-b196-e00e7f669ec9" [ 1226.276122] env[65522]: _type = "Task" [ 1226.276122] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.284814] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8bc017-1452-4bc4-90d6-f042fbc9d8aa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.294332] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e4e8fd5f-0bc1-437d-a8db-457c5ea05344" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1226.294639] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Processing image e4e8fd5f-0bc1-437d-a8db-457c5ea05344 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1226.294915] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e4e8fd5f-0bc1-437d-a8db-457c5ea05344/e4e8fd5f-0bc1-437d-a8db-457c5ea05344.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1226.295221] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e4e8fd5f-0bc1-437d-a8db-457c5ea05344/e4e8fd5f-0bc1-437d-a8db-457c5ea05344.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1226.295426] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1226.295807] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-03cdf3bd-f05a-4347-bb69-081ea5d383a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.307321] env[65522]: DEBUG nova.compute.provider_tree [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1226.347936] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1226.348365] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1226.350229] env[65522]: WARNING neutronclient.v2_0.client [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1226.351060] env[65522]: WARNING openstack [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1226.351581] env[65522]: WARNING openstack [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1226.360844] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc8a5f94-7e28-4fdc-9610-2144e690731d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.370932] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1226.370932] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52821744-a888-2363-ef8d-e097f0586563" [ 1226.370932] env[65522]: _type = "Task" [ 1226.370932] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.385766] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115051, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.389646] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52821744-a888-2363-ef8d-e097f0586563, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.460968] env[65522]: DEBUG nova.network.neutron [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Updated VIF entry in instance network info cache for port fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1226.461366] env[65522]: DEBUG nova.network.neutron [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Updating instance_info_cache with network_info: [{"id": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "address": "fa:16:3e:c3:dd:ad", "network": {"id": "f36c7616-6aee-4137-8f00-350aac5cb938", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1830347608-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1546bbdca07c443d84abea1155cfde37", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4c5eb94-841c-4713-985a-8fc4117fbaf1", "external-id": "nsx-vlan-transportzone-425", "segmentation_id": 425, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfed73ed8-c0", "ovs_interfaceid": "fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1226.685620] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115053, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.815813] env[65522]: DEBUG nova.scheduler.client.report [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1226.888091] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115051, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.586862} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.891147] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 874813b4-d0d6-4559-82e2-9fa2ea4feb15/874813b4-d0d6-4559-82e2-9fa2ea4feb15.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1226.891430] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1226.891883] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d55ed4d-91e0-467f-a684-c6a06e70e1a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.894219] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Preparing fetch location {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1226.894492] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Fetch image to [datastore2] OSTACK_IMG_f0aba43c-d831-4114-9bea-1728852871c6/OSTACK_IMG_f0aba43c-d831-4114-9bea-1728852871c6.vmdk {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1226.894702] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Downloading stream optimized image e4e8fd5f-0bc1-437d-a8db-457c5ea05344 to [datastore2] OSTACK_IMG_f0aba43c-d831-4114-9bea-1728852871c6/OSTACK_IMG_f0aba43c-d831-4114-9bea-1728852871c6.vmdk on the data store datastore2 as vApp {{(pid=65522) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1226.894842] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Downloading image file data e4e8fd5f-0bc1-437d-a8db-457c5ea05344 to the ESX as VM named 'OSTACK_IMG_f0aba43c-d831-4114-9bea-1728852871c6' {{(pid=65522) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1226.905657] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1226.905657] env[65522]: value = "task-5115054" [ 1226.905657] env[65522]: _type = "Task" [ 1226.905657] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.922331] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115054, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.964783] env[65522]: DEBUG oslo_concurrency.lockutils [req-ce6c9997-ea52-4372-832f-b20e23f39735 req-d1d3481f-34aa-4742-a123-a952b0580cde service nova] Releasing lock "refresh_cache-3b82f9ce-ea59-430d-adb6-918cc6cc48a4" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1226.986628] env[65522]: DEBUG oslo_vmware.rw_handles [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1226.986628] env[65522]: value = "resgroup-9" [ 1226.986628] env[65522]: _type = "ResourcePool" [ 1226.986628] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1226.987021] env[65522]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-2e314b02-24bc-496b-ad45-f193316df0f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.015230] env[65522]: DEBUG oslo_vmware.rw_handles [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lease: (returnval){ [ 1227.015230] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cf7b82-d752-5048-482d-aeb97f6d0db5" [ 1227.015230] env[65522]: _type = "HttpNfcLease" [ 1227.015230] env[65522]: } obtained for vApp import into resource pool (val){ [ 1227.015230] env[65522]: value = "resgroup-9" [ 1227.015230] env[65522]: _type = "ResourcePool" [ 1227.015230] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1227.015851] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the lease: (returnval){ [ 1227.015851] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cf7b82-d752-5048-482d-aeb97f6d0db5" [ 1227.015851] env[65522]: _type = "HttpNfcLease" [ 1227.015851] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1227.027731] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1227.027731] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cf7b82-d752-5048-482d-aeb97f6d0db5" [ 1227.027731] env[65522]: _type = "HttpNfcLease" [ 1227.027731] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1227.186047] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115053, 'name': MoveVirtualDisk_Task} progress is 26%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.282057] env[65522]: DEBUG oslo_concurrency.lockutils [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "3987d71f-47ab-4dd4-8426-95d78503effa" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1227.282407] env[65522]: DEBUG oslo_concurrency.lockutils [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1227.321552] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.361s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1227.321823] env[65522]: INFO nova.compute.manager [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Migrating [ 1227.330901] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.258s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1227.420997] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115054, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085593} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.421314] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1227.422642] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-641db9e2-8871-4f4b-92a4-7ec9fbd4272b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.451940] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] 874813b4-d0d6-4559-82e2-9fa2ea4feb15/874813b4-d0d6-4559-82e2-9fa2ea4feb15.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1227.452438] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1227.452677] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1227.452894] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1227.453091] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1227.453427] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1227.455198] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86d91392-952b-4c19-82c8-e8d4ed0fdd1e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.473114] env[65522]: INFO nova.compute.manager [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Terminating instance [ 1227.486351] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1227.486351] env[65522]: value = "task-5115056" [ 1227.486351] env[65522]: _type = "Task" [ 1227.486351] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.505794] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115056, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.526488] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1227.526488] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cf7b82-d752-5048-482d-aeb97f6d0db5" [ 1227.526488] env[65522]: _type = "HttpNfcLease" [ 1227.526488] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1227.685496] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115053, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.786209] env[65522]: DEBUG nova.compute.utils [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1227.846596] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1227.847037] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1227.847387] env[65522]: DEBUG nova.network.neutron [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1227.979717] env[65522]: DEBUG nova.compute.manager [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1227.980386] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1227.981103] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffdbaf21-34d8-41fa-bb67-76998f57321b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.996195] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1227.997118] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e86d26b0-2e01-4e81-99b2-e8d5c36da6e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.005471] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115056, 'name': ReconfigVM_Task, 'duration_secs': 0.3749} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.007245] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Reconfigured VM instance instance-00000076 to attach disk [datastore2] 874813b4-d0d6-4559-82e2-9fa2ea4feb15/874813b4-d0d6-4559-82e2-9fa2ea4feb15.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1228.008116] env[65522]: DEBUG oslo_vmware.api [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1228.008116] env[65522]: value = "task-5115057" [ 1228.008116] env[65522]: _type = "Task" [ 1228.008116] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.008502] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-caefe4b0-4cca-42d0-95f3-b3ce8d6aa7a6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.026547] env[65522]: DEBUG oslo_vmware.api [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115057, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.026872] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1228.026872] env[65522]: value = "task-5115058" [ 1228.026872] env[65522]: _type = "Task" [ 1228.026872] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.034123] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1228.034123] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cf7b82-d752-5048-482d-aeb97f6d0db5" [ 1228.034123] env[65522]: _type = "HttpNfcLease" [ 1228.034123] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1228.034506] env[65522]: DEBUG oslo_vmware.rw_handles [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1228.034506] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cf7b82-d752-5048-482d-aeb97f6d0db5" [ 1228.034506] env[65522]: _type = "HttpNfcLease" [ 1228.034506] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1228.035430] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d68940-2cb7-44cf-9da9-3170a7bbbd6e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.042535] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115058, 'name': Rename_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.052556] env[65522]: DEBUG oslo_vmware.rw_handles [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e2be9a-a6d4-a604-08d0-463c26f1856b/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1228.053208] env[65522]: DEBUG oslo_vmware.rw_handles [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e2be9a-a6d4-a604-08d0-463c26f1856b/disk-0.vmdk. {{(pid=65522) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1228.121903] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fa2b2843-4611-43d5-888b-b2a9f3d1adc6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.189567] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115053, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.290226] env[65522]: DEBUG oslo_concurrency.lockutils [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1228.351411] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Applying migration context for instance daffdd40-9ee3-4a80-b670-5772395a32ae as it has an incoming, in-progress migration 5817abdf-1156-4dda-8607-b249c325820b. Migration status is pre-migrating {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1228.353357] env[65522]: INFO nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating resource usage from migration 5817abdf-1156-4dda-8607-b249c325820b [ 1228.359823] env[65522]: WARNING neutronclient.v2_0.client [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1228.360475] env[65522]: WARNING openstack [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1228.360852] env[65522]: WARNING openstack [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1228.382223] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7617189c-a902-42e7-8165-0e7c4a1de06d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1228.382387] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance aebb7082-f88c-42f3-9d9f-93cb2a6f545b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1228.382495] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 3987d71f-47ab-4dd4-8426-95d78503effa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1228.382615] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 874b6a07-ebba-4f75-937e-2281af1598b8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1228.382764] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 3adfefa5-fb19-44b2-b3c5-42f2e2918673 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1228.382888] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance ae3455d8-2f38-42ad-b16d-d98aef92b2a8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1228.382998] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 1b0fcdbb-aa9e-4915-a8c6-1167aa888390 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1228.383126] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 2981b635-43c8-4bd6-9991-e6af0be82f3c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1228.383252] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 3b82f9ce-ea59-430d-adb6-918cc6cc48a4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1228.383376] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 874813b4-d0d6-4559-82e2-9fa2ea4feb15 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1228.383490] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Migration 5817abdf-1156-4dda-8607-b249c325820b is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1228.383659] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance daffdd40-9ee3-4a80-b670-5772395a32ae actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1228.383877] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1228.384142] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2944MB phys_disk=100GB used_disk=10GB total_vcpus=48 used_vcpus=12 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '11', 'num_vm_active': '7', 'num_task_None': '7', 'num_os_type_None': '11', 'num_proj_91597233ae9c44c094f4c32d90332fa6': '3', 'io_workload': '2', 'num_vm_shelved_offloaded': '2', 'num_task_spawning': '3', 'num_proj_fdba2dbc660145b7a98af8bd3f77a035': '1', 'num_proj_1546bbdca07c443d84abea1155cfde37': '1', 'num_proj_a80f979f3dc0477e9462b47f7aa87f14': '1', 'num_proj_f93c790ff61543bd8e134bcf9cb20bb2': '1', 'num_vm_rescued': '1', 'num_proj_15229af4da7a4ff9963369df50fb9d5b': '2', 'num_proj_0d1239b79ae94cceb89ae7a8bd57da08': '1', 'num_task_resize_prep': '1', 'num_proj_fa11b46d9fe144f391233e6eb9c819d7': '1', 'num_vm_building': '1'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1228.519373] env[65522]: WARNING openstack [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1228.519818] env[65522]: WARNING openstack [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1228.544581] env[65522]: DEBUG oslo_vmware.api [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115057, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.552689] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115058, 'name': Rename_Task, 'duration_secs': 0.187126} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.556242] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1228.559770] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-134a1cc9-b633-47d2-b768-f3923468698c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.572207] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1228.572207] env[65522]: value = "task-5115059" [ 1228.572207] env[65522]: _type = "Task" [ 1228.572207] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.592737] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115059, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.622448] env[65522]: WARNING neutronclient.v2_0.client [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1228.623154] env[65522]: WARNING openstack [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1228.623535] env[65522]: WARNING openstack [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1228.635296] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4001e9eb-2fd2-4002-b940-870c3b4c5da8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.646454] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6419bc13-024e-4cf0-beba-93e353fed8f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.702647] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8659eb3c-79a2-4ee4-b9c8-a4f25010dd67 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.715578] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115053, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.721457] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cfa45f5-1ad6-4c88-9668-6a3c5ac33bc4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.741018] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1228.750875] env[65522]: DEBUG nova.network.neutron [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating instance_info_cache with network_info: [{"id": "6b80135d-6228-4366-83aa-ff70808ae005", "address": "fa:16:3e:82:79:a0", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b80135d-62", "ovs_interfaceid": "6b80135d-6228-4366-83aa-ff70808ae005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1229.023956] env[65522]: DEBUG oslo_vmware.api [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115057, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.085174] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115059, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.199368] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115053, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.951049} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.199539] env[65522]: INFO nova.virt.vmwareapi.ds_util [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_cab48bd4-4166-4507-af63-a2c2a1d58384/OSTACK_IMG_cab48bd4-4166-4507-af63-a2c2a1d58384.vmdk to [datastore2] devstack-image-cache_base/7832ad6e-429a-4894-9720-58fd3520c4e2/7832ad6e-429a-4894-9720-58fd3520c4e2.vmdk. [ 1229.199775] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Cleaning up location [datastore2] OSTACK_IMG_cab48bd4-4166-4507-af63-a2c2a1d58384 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1229.199978] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_cab48bd4-4166-4507-af63-a2c2a1d58384 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1229.200311] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9154bcd5-af19-4cad-8f93-99e1a2eb8da2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.209934] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1229.209934] env[65522]: value = "task-5115060" [ 1229.209934] env[65522]: _type = "Task" [ 1229.209934] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.220541] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115060, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.244791] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1229.254324] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1229.369739] env[65522]: DEBUG oslo_concurrency.lockutils [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "3987d71f-47ab-4dd4-8426-95d78503effa" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1229.369945] env[65522]: DEBUG oslo_concurrency.lockutils [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1229.370211] env[65522]: INFO nova.compute.manager [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Attaching volume 43d813d9-8062-43f5-8ca7-9f86fdd62d1c to /dev/sdb [ 1229.401993] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a06c951-53d7-4cd2-a0e0-0416a2f4c48b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.409673] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-389833af-ba24-47fe-81e3-5b99ecb6fd1b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.424825] env[65522]: DEBUG nova.virt.block_device [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Updating existing volume attachment record: 1c70306c-97bb-4e9e-a7ea-8d3e115ac0fb {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1229.524195] env[65522]: DEBUG oslo_vmware.api [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115057, 'name': PowerOffVM_Task, 'duration_secs': 1.060213} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.524944] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1229.525146] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1229.525425] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-60b56018-3431-4b2c-878d-12e818045136 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.583855] env[65522]: DEBUG oslo_vmware.api [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115059, 'name': PowerOnVM_Task, 'duration_secs': 0.567195} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.584154] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1229.584358] env[65522]: INFO nova.compute.manager [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Took 8.25 seconds to spawn the instance on the hypervisor. [ 1229.584537] env[65522]: DEBUG nova.compute.manager [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1229.585372] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aea028f-8212-4d16-8538-ffd782ba8e4f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.603709] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1229.603979] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1229.604180] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleting the datastore file [datastore2] ae3455d8-2f38-42ad-b16d-d98aef92b2a8 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1229.604522] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c2867ac-8e08-4564-a16f-920b4d4c896d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.611927] env[65522]: DEBUG oslo_vmware.api [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for the task: (returnval){ [ 1229.611927] env[65522]: value = "task-5115063" [ 1229.611927] env[65522]: _type = "Task" [ 1229.611927] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.621566] env[65522]: DEBUG oslo_vmware.api [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115063, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.738918] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115060, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101914} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.743629] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1229.744031] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7832ad6e-429a-4894-9720-58fd3520c4e2/7832ad6e-429a-4894-9720-58fd3520c4e2.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1229.744466] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7832ad6e-429a-4894-9720-58fd3520c4e2/7832ad6e-429a-4894-9720-58fd3520c4e2.vmdk to [datastore2] 2981b635-43c8-4bd6-9991-e6af0be82f3c/2981b635-43c8-4bd6-9991-e6af0be82f3c.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1229.744824] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bda168eb-be3b-49c1-8532-de489a1bbfc1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.750523] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1229.750832] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.420s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1229.753454] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1229.753454] env[65522]: value = "task-5115064" [ 1229.753454] env[65522]: _type = "Task" [ 1229.753454] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.770578] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115064, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.879243] env[65522]: DEBUG oslo_vmware.rw_handles [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Completed reading data from the image iterator. {{(pid=65522) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1229.879515] env[65522]: DEBUG oslo_vmware.rw_handles [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e2be9a-a6d4-a604-08d0-463c26f1856b/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1229.880610] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ee4c1d-0b14-4227-837e-5f16bb8f2931 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.888892] env[65522]: DEBUG oslo_vmware.rw_handles [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e2be9a-a6d4-a604-08d0-463c26f1856b/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1229.889082] env[65522]: DEBUG oslo_vmware.rw_handles [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e2be9a-a6d4-a604-08d0-463c26f1856b/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1229.889400] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-e59aa4ac-bcec-4973-9b86-5ccc0202ff9a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.105431] env[65522]: INFO nova.compute.manager [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Took 15.53 seconds to build instance. [ 1230.123760] env[65522]: DEBUG oslo_vmware.api [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Task: {'id': task-5115063, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212834} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.125048] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1230.125435] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1230.125672] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1230.125919] env[65522]: INFO nova.compute.manager [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Took 2.15 seconds to destroy the instance on the hypervisor. [ 1230.126287] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1230.126919] env[65522]: DEBUG nova.compute.manager [-] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1230.127098] env[65522]: DEBUG nova.network.neutron [-] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1230.127467] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1230.128116] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1230.128438] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1230.163487] env[65522]: DEBUG oslo_vmware.rw_handles [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e2be9a-a6d4-a604-08d0-463c26f1856b/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1230.163828] env[65522]: INFO nova.virt.vmwareapi.images [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Downloaded image file data e4e8fd5f-0bc1-437d-a8db-457c5ea05344 [ 1230.164864] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-807548b8-c7d4-463c-aca1-9e406eb7157c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.173591] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1230.187169] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1868ae82-390e-4a42-837f-9209f96ce593 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.232567] env[65522]: INFO nova.virt.vmwareapi.images [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] The imported VM was unregistered [ 1230.236194] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Caching image {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1230.236299] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Creating directory with path [datastore2] devstack-image-cache_base/e4e8fd5f-0bc1-437d-a8db-457c5ea05344 {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1230.236698] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7d1b08a8-ee86-4cf9-a098-9e76207201cd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.272686] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115064, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.296090] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Created directory with path [datastore2] devstack-image-cache_base/e4e8fd5f-0bc1-437d-a8db-457c5ea05344 {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1230.296993] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_f0aba43c-d831-4114-9bea-1728852871c6/OSTACK_IMG_f0aba43c-d831-4114-9bea-1728852871c6.vmdk to [datastore2] devstack-image-cache_base/e4e8fd5f-0bc1-437d-a8db-457c5ea05344/e4e8fd5f-0bc1-437d-a8db-457c5ea05344.vmdk. {{(pid=65522) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1230.296993] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-46fa71ab-45ac-4bd8-843c-5c7298aef219 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.310543] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1230.310543] env[65522]: value = "task-5115066" [ 1230.310543] env[65522]: _type = "Task" [ 1230.310543] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.327724] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115066, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.493158] env[65522]: DEBUG nova.compute.manager [req-6c0cc07f-c78c-4d02-b41b-33b831065f22 req-e4a32ca0-e474-4f60-b0d1-ff101f8020bd service nova] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Received event network-vif-deleted-499e91eb-948a-4b54-b7a6-ce30af8315ea {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1230.493356] env[65522]: INFO nova.compute.manager [req-6c0cc07f-c78c-4d02-b41b-33b831065f22 req-e4a32ca0-e474-4f60-b0d1-ff101f8020bd service nova] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Neutron deleted interface 499e91eb-948a-4b54-b7a6-ce30af8315ea; detaching it from the instance and deleting it from the info cache [ 1230.493596] env[65522]: DEBUG nova.network.neutron [req-6c0cc07f-c78c-4d02-b41b-33b831065f22 req-e4a32ca0-e474-4f60-b0d1-ff101f8020bd service nova] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1230.608789] env[65522]: DEBUG oslo_concurrency.lockutils [None req-dc8f4e1a-8e5b-4b0f-8af2-b59d39b2835d tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.043s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1230.609208] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 9.920s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1230.609408] env[65522]: INFO nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] During sync_power_state the instance has a pending task (networking). Skip. [ 1230.609653] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1230.725897] env[65522]: INFO nova.compute.manager [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Rescuing [ 1230.726248] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "refresh_cache-874813b4-d0d6-4559-82e2-9fa2ea4feb15" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1230.726410] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquired lock "refresh_cache-874813b4-d0d6-4559-82e2-9fa2ea4feb15" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1230.726580] env[65522]: DEBUG nova.network.neutron [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1230.770160] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115064, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.778367] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c315040-0e32-483a-b29e-866aed1fa3de {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.804644] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating instance 'daffdd40-9ee3-4a80-b670-5772395a32ae' progress to 0 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1230.823228] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115066, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.963795] env[65522]: DEBUG nova.network.neutron [-] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1230.996378] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-717ce80a-d1dc-461b-beaf-f8e0abfc01c4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.012933] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f519d7e3-18fe-4ed0-abd3-b93303f37ff2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.058870] env[65522]: DEBUG nova.compute.manager [req-6c0cc07f-c78c-4d02-b41b-33b831065f22 req-e4a32ca0-e474-4f60-b0d1-ff101f8020bd service nova] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Detach interface failed, port_id=499e91eb-948a-4b54-b7a6-ce30af8315ea, reason: Instance ae3455d8-2f38-42ad-b16d-d98aef92b2a8 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1231.229934] env[65522]: WARNING neutronclient.v2_0.client [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1231.230691] env[65522]: WARNING openstack [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1231.231226] env[65522]: WARNING openstack [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1231.268729] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115064, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.312110] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1231.312507] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb1f6a6a-9821-4327-8732-00066798f2b7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.331938] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115066, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.334686] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1231.334686] env[65522]: value = "task-5115067" [ 1231.334686] env[65522]: _type = "Task" [ 1231.334686] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.349619] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115067, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.355675] env[65522]: WARNING openstack [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1231.356097] env[65522]: WARNING openstack [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1231.413972] env[65522]: WARNING neutronclient.v2_0.client [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1231.414745] env[65522]: WARNING openstack [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1231.415156] env[65522]: WARNING openstack [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1231.469812] env[65522]: INFO nova.compute.manager [-] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Took 1.34 seconds to deallocate network for instance. [ 1231.520092] env[65522]: DEBUG nova.network.neutron [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Updating instance_info_cache with network_info: [{"id": "6ff7afa0-2b0e-450c-a755-77a3241a2527", "address": "fa:16:3e:c7:7a:9b", "network": {"id": "26dda233-0dd0-40ae-b56f-c84b1b91e92a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1197499239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "15229af4da7a4ff9963369df50fb9d5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ff7afa0-2b", "ovs_interfaceid": "6ff7afa0-2b0e-450c-a755-77a3241a2527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1231.771060] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115064, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.825502] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115066, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.846948] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115067, 'name': PowerOffVM_Task, 'duration_secs': 0.222702} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.848117] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1231.848117] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating instance 'daffdd40-9ee3-4a80-b670-5772395a32ae' progress to 17 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1231.979908] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1231.980186] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1231.980408] env[65522]: DEBUG nova.objects.instance [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lazy-loading 'resources' on Instance uuid ae3455d8-2f38-42ad-b16d-d98aef92b2a8 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1232.023905] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Releasing lock "refresh_cache-874813b4-d0d6-4559-82e2-9fa2ea4feb15" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1232.270623] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115064, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.328341] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115066, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.354509] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1232.354806] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1232.355047] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1232.355265] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1232.355419] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1232.355566] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1232.355782] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1232.355949] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1232.356141] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1232.356307] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1232.356483] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1232.362516] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ff77ba9-2530-40aa-983e-3937231477fb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.381548] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1232.381548] env[65522]: value = "task-5115069" [ 1232.381548] env[65522]: _type = "Task" [ 1232.381548] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.393614] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115069, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.696392] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4baee8e3-c845-4b4a-b0ed-834f0caa7115 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.708670] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a094f182-48b6-44d1-8548-f89b502b5d22 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.745112] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d0d78d-65fd-4793-97b6-a9bae72f243c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.756028] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b0dfb5-70c4-4c1a-acb9-8c9355604c64 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.771077] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115064, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.652277} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.779582] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7832ad6e-429a-4894-9720-58fd3520c4e2/7832ad6e-429a-4894-9720-58fd3520c4e2.vmdk to [datastore2] 2981b635-43c8-4bd6-9991-e6af0be82f3c/2981b635-43c8-4bd6-9991-e6af0be82f3c.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1232.780215] env[65522]: DEBUG nova.compute.provider_tree [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1232.782958] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0267e0c1-3319-412e-8842-f6f6eefebcdb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.812083] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 2981b635-43c8-4bd6-9991-e6af0be82f3c/2981b635-43c8-4bd6-9991-e6af0be82f3c.vmdk or device None with type streamOptimized {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1232.812877] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19294d36-e78b-4e35-a88f-2a9180fc26e2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.841819] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115066, 'name': MoveVirtualDisk_Task} progress is 26%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.843247] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1232.843247] env[65522]: value = "task-5115070" [ 1232.843247] env[65522]: _type = "Task" [ 1232.843247] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.854398] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115070, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.895357] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115069, 'name': ReconfigVM_Task, 'duration_secs': 0.269676} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.895695] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating instance 'daffdd40-9ee3-4a80-b670-5772395a32ae' progress to 33 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1233.284509] env[65522]: DEBUG nova.scheduler.client.report [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1233.343566] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115066, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.355958] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115070, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.403652] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1233.403974] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1233.404131] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1233.404375] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1233.404635] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1233.404858] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1233.405151] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1233.405380] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1233.405618] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1233.405761] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1233.406040] env[65522]: DEBUG nova.virt.hardware [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1233.412322] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Reconfiguring VM instance instance-00000074 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1233.412703] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b52c0d16-839c-4c62-8f7c-1fa9ef994820 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.441384] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1233.441384] env[65522]: value = "task-5115071" [ 1233.441384] env[65522]: _type = "Task" [ 1233.441384] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.455165] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115071, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.573052] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1233.573494] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ab2c70a4-4e92-4b84-aa8b-894b626b3b30 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.584839] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1233.584839] env[65522]: value = "task-5115072" [ 1233.584839] env[65522]: _type = "Task" [ 1233.584839] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.596948] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115072, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.791607] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.811s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1233.818480] env[65522]: INFO nova.scheduler.client.report [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Deleted allocations for instance ae3455d8-2f38-42ad-b16d-d98aef92b2a8 [ 1233.846077] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115066, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.858156] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115070, 'name': ReconfigVM_Task, 'duration_secs': 0.731017} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.858466] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 2981b635-43c8-4bd6-9991-e6af0be82f3c/2981b635-43c8-4bd6-9991-e6af0be82f3c.vmdk or device None with type streamOptimized {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1233.859385] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-872419af-9ba4-4001-9759-0fbdb5ca4715 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.869477] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1233.869477] env[65522]: value = "task-5115073" [ 1233.869477] env[65522]: _type = "Task" [ 1233.869477] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.881917] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115073, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.954802] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115071, 'name': ReconfigVM_Task, 'duration_secs': 0.279864} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.954802] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Reconfigured VM instance instance-00000074 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1233.955545] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f3dfe3-2451-4578-b1c7-5684e7781e4a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.983700] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] volume-eacd0d18-4608-43ee-a54c-6655db952ccb/volume-eacd0d18-4608-43ee-a54c-6655db952ccb.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1233.985122] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1233.985353] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994975', 'volume_id': '43d813d9-8062-43f5-8ca7-9f86fdd62d1c', 'name': 'volume-43d813d9-8062-43f5-8ca7-9f86fdd62d1c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3987d71f-47ab-4dd4-8426-95d78503effa', 'attached_at': '', 'detached_at': '', 'volume_id': '43d813d9-8062-43f5-8ca7-9f86fdd62d1c', 'serial': '43d813d9-8062-43f5-8ca7-9f86fdd62d1c'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1233.985668] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e118660d-ab82-4b93-889f-b2c7ffdb0d73 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.001148] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44e4955-4813-4a67-9ecc-cd62d2eee90a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.022753] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fac46e3-94b3-45ea-8798-e62e2a45026e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.026058] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1234.026058] env[65522]: value = "task-5115074" [ 1234.026058] env[65522]: _type = "Task" [ 1234.026058] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.055851] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] volume-43d813d9-8062-43f5-8ca7-9f86fdd62d1c/volume-43d813d9-8062-43f5-8ca7-9f86fdd62d1c.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1234.060185] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8df6ce8-fefb-4e52-8fc5-e9c00360732e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.074430] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115074, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.081788] env[65522]: DEBUG oslo_vmware.api [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1234.081788] env[65522]: value = "task-5115075" [ 1234.081788] env[65522]: _type = "Task" [ 1234.081788] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.097298] env[65522]: DEBUG oslo_vmware.api [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115075, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.102136] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115072, 'name': PowerOffVM_Task, 'duration_secs': 0.320858} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.102584] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1234.103877] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3f00ba-53d8-4d8c-9127-f835049bb0a1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.131582] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e403a2-732f-4e14-9d79-719f4c60a108 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.176926] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1234.177306] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d675ce5-6e8b-44d2-b118-7348deb665b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.188327] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1234.188327] env[65522]: value = "task-5115076" [ 1234.188327] env[65522]: _type = "Task" [ 1234.188327] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.202213] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1234.202520] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1234.202860] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.203152] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1234.203387] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1234.203847] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bed834c9-5b28-4f53-bf37-f04343977781 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.225107] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1234.225397] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1234.226317] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e4ffb32-147f-46a3-ba31-65ce00c2ca19 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.235485] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1234.235485] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c8cc42-4f72-bea1-f58c-bba2ca8d1abe" [ 1234.235485] env[65522]: _type = "Task" [ 1234.235485] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.249257] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c8cc42-4f72-bea1-f58c-bba2ca8d1abe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.328083] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8fae9215-32e6-4996-a052-73e9f4c45be5 tempest-DeleteServersTestJSON-868563199 tempest-DeleteServersTestJSON-868563199-project-member] Lock "ae3455d8-2f38-42ad-b16d-d98aef92b2a8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.875s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1234.345440] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115066, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.382773] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115073, 'name': Rename_Task, 'duration_secs': 0.276147} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.383228] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1234.383575] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-48837a5f-883a-4252-9ec3-6aa9dd3ae667 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.394516] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1234.394516] env[65522]: value = "task-5115077" [ 1234.394516] env[65522]: _type = "Task" [ 1234.394516] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.409184] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115077, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.540030] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115074, 'name': ReconfigVM_Task, 'duration_secs': 0.44344} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.540534] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Reconfigured VM instance instance-00000074 to attach disk [datastore1] volume-eacd0d18-4608-43ee-a54c-6655db952ccb/volume-eacd0d18-4608-43ee-a54c-6655db952ccb.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1234.540874] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating instance 'daffdd40-9ee3-4a80-b670-5772395a32ae' progress to 50 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1234.594548] env[65522]: DEBUG oslo_vmware.api [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115075, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.746704] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c8cc42-4f72-bea1-f58c-bba2ca8d1abe, 'name': SearchDatastore_Task, 'duration_secs': 0.088189} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.747569] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0c9a8be-8387-40e2-948d-50f368368083 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.754559] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1234.754559] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f8f8b5-9bc3-e4f4-a738-14d90266a9fa" [ 1234.754559] env[65522]: _type = "Task" [ 1234.754559] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.763268] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f8f8b5-9bc3-e4f4-a738-14d90266a9fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.842321] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115066, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.905238] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115077, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.048600] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d0501a-cf93-4748-a708-1a28a54ffe51 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.070815] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59887a39-76d9-4607-bba1-ab26b81bd244 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.089412] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating instance 'daffdd40-9ee3-4a80-b670-5772395a32ae' progress to 67 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1235.108746] env[65522]: DEBUG oslo_vmware.api [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115075, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.269458] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f8f8b5-9bc3-e4f4-a738-14d90266a9fa, 'name': SearchDatastore_Task, 'duration_secs': 0.070152} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.269773] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1235.270141] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 874813b4-d0d6-4559-82e2-9fa2ea4feb15/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk. {{(pid=65522) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1235.270449] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d65b2dda-28c3-496d-a5c2-2815d804d5e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.279218] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1235.279218] env[65522]: value = "task-5115079" [ 1235.279218] env[65522]: _type = "Task" [ 1235.279218] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.293743] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115079, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.345492] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115066, 'name': MoveVirtualDisk_Task, 'duration_secs': 4.735468} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.345843] env[65522]: INFO nova.virt.vmwareapi.ds_util [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_f0aba43c-d831-4114-9bea-1728852871c6/OSTACK_IMG_f0aba43c-d831-4114-9bea-1728852871c6.vmdk to [datastore2] devstack-image-cache_base/e4e8fd5f-0bc1-437d-a8db-457c5ea05344/e4e8fd5f-0bc1-437d-a8db-457c5ea05344.vmdk. [ 1235.345978] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Cleaning up location [datastore2] OSTACK_IMG_f0aba43c-d831-4114-9bea-1728852871c6 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1235.346169] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_f0aba43c-d831-4114-9bea-1728852871c6 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1235.346452] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b0eb5f21-c8ff-4b86-9324-a8d2f5ab498d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.354230] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1235.354230] env[65522]: value = "task-5115080" [ 1235.354230] env[65522]: _type = "Task" [ 1235.354230] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.363124] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115080, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.406100] env[65522]: DEBUG oslo_vmware.api [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115077, 'name': PowerOnVM_Task, 'duration_secs': 0.679929} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.406368] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1235.512829] env[65522]: DEBUG nova.compute.manager [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1235.513954] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb9ba71-0a31-4c08-ab5e-6326c83b037d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.603086] env[65522]: DEBUG oslo_vmware.api [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115075, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.791295] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115079, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.865995] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115080, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.040624} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.866350] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1235.866522] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e4e8fd5f-0bc1-437d-a8db-457c5ea05344/e4e8fd5f-0bc1-437d-a8db-457c5ea05344.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1235.866784] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e4e8fd5f-0bc1-437d-a8db-457c5ea05344/e4e8fd5f-0bc1-437d-a8db-457c5ea05344.vmdk to [datastore2] 3b82f9ce-ea59-430d-adb6-918cc6cc48a4/3b82f9ce-ea59-430d-adb6-918cc6cc48a4.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1235.867085] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9ce7d944-a7a3-4f1d-9349-1b67e55b24e1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.876503] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1235.876503] env[65522]: value = "task-5115081" [ 1235.876503] env[65522]: _type = "Task" [ 1235.876503] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.886689] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115081, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.033563] env[65522]: DEBUG oslo_concurrency.lockutils [None req-69eb057e-7b92-4f4d-919c-2acb4efb6c00 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 24.031s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1236.035069] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 15.349s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1236.035069] env[65522]: INFO nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] During sync_power_state the instance has a pending task (spawning). Skip. [ 1236.035069] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1236.100067] env[65522]: DEBUG oslo_vmware.api [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115075, 'name': ReconfigVM_Task, 'duration_secs': 1.541059} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.100428] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Reconfigured VM instance instance-00000070 to attach disk [datastore1] volume-43d813d9-8062-43f5-8ca7-9f86fdd62d1c/volume-43d813d9-8062-43f5-8ca7-9f86fdd62d1c.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1236.105563] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59c294f0-4632-4aec-968b-3e5705ecc421 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.119969] env[65522]: DEBUG oslo_vmware.api [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1236.119969] env[65522]: value = "task-5115082" [ 1236.119969] env[65522]: _type = "Task" [ 1236.119969] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.128558] env[65522]: DEBUG oslo_vmware.api [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115082, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.295056] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115079, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.770543} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.295352] env[65522]: INFO nova.virt.vmwareapi.ds_util [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] 874813b4-d0d6-4559-82e2-9fa2ea4feb15/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk. [ 1236.296360] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad8fd5d-ed5d-4a29-85f2-d325382e7bdb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.325975] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] 874813b4-d0d6-4559-82e2-9fa2ea4feb15/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1236.326373] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f7085e4-b796-41c3-928f-f4775247891c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.348672] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1236.348672] env[65522]: value = "task-5115083" [ 1236.348672] env[65522]: _type = "Task" [ 1236.348672] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.360716] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115083, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.388233] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115081, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.631455] env[65522]: DEBUG oslo_vmware.api [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115082, 'name': ReconfigVM_Task, 'duration_secs': 0.173295} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.631809] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994975', 'volume_id': '43d813d9-8062-43f5-8ca7-9f86fdd62d1c', 'name': 'volume-43d813d9-8062-43f5-8ca7-9f86fdd62d1c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3987d71f-47ab-4dd4-8426-95d78503effa', 'attached_at': '', 'detached_at': '', 'volume_id': '43d813d9-8062-43f5-8ca7-9f86fdd62d1c', 'serial': '43d813d9-8062-43f5-8ca7-9f86fdd62d1c'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1236.712038] env[65522]: WARNING neutronclient.v2_0.client [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1236.753433] env[65522]: DEBUG nova.network.neutron [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Port 6b80135d-6228-4366-83aa-ff70808ae005 binding to destination host cpu-1 is already ACTIVE {{(pid=65522) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 1236.863383] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115083, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.890363] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115081, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.363639] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115083, 'name': ReconfigVM_Task, 'duration_secs': 0.670393} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.364029] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Reconfigured VM instance instance-00000076 to attach disk [datastore2] 874813b4-d0d6-4559-82e2-9fa2ea4feb15/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1237.365293] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c28d1f-6b3e-4225-98c2-c96e15a79117 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.398060] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31f2a881-9a58-4cb5-a15b-b761e2376c54 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.418413] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115081, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.419622] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1237.419622] env[65522]: value = "task-5115084" [ 1237.419622] env[65522]: _type = "Task" [ 1237.419622] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.429072] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115084, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.677945] env[65522]: DEBUG nova.objects.instance [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lazy-loading 'flavor' on Instance uuid 3987d71f-47ab-4dd4-8426-95d78503effa {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1237.754679] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c291e3-888f-4d07-b8ad-096a6b5f4b2d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.770881] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c9db8083-60e3-4aa4-abd3-418d2acbc584 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Suspending the VM {{(pid=65522) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1237.772229] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-163e715c-39bc-4928-bdf1-6b441bde8ad2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.782874] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "daffdd40-9ee3-4a80-b670-5772395a32ae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1237.782874] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "daffdd40-9ee3-4a80-b670-5772395a32ae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1237.782874] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "daffdd40-9ee3-4a80-b670-5772395a32ae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1237.791615] env[65522]: DEBUG oslo_vmware.api [None req-c9db8083-60e3-4aa4-abd3-418d2acbc584 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1237.791615] env[65522]: value = "task-5115085" [ 1237.791615] env[65522]: _type = "Task" [ 1237.791615] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.800735] env[65522]: DEBUG oslo_vmware.api [None req-c9db8083-60e3-4aa4-abd3-418d2acbc584 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115085, 'name': SuspendVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.900695] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115081, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.930908] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115084, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.184257] env[65522]: DEBUG oslo_concurrency.lockutils [None req-893aed0a-9472-4819-9fd9-5838b71ded5e tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.814s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1238.302022] env[65522]: DEBUG oslo_vmware.api [None req-c9db8083-60e3-4aa4-abd3-418d2acbc584 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115085, 'name': SuspendVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.406052] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115081, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.455979} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.406417] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e4e8fd5f-0bc1-437d-a8db-457c5ea05344/e4e8fd5f-0bc1-437d-a8db-457c5ea05344.vmdk to [datastore2] 3b82f9ce-ea59-430d-adb6-918cc6cc48a4/3b82f9ce-ea59-430d-adb6-918cc6cc48a4.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1238.407084] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1beae16-5951-45bc-9414-5bf47b62f987 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.430166] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 3b82f9ce-ea59-430d-adb6-918cc6cc48a4/3b82f9ce-ea59-430d-adb6-918cc6cc48a4.vmdk or device None with type streamOptimized {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1238.434286] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7acc2028-51e8-44ec-ba65-b2f5880d6e26 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.454851] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115084, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.456316] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1238.456316] env[65522]: value = "task-5115086" [ 1238.456316] env[65522]: _type = "Task" [ 1238.456316] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.464592] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115086, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.602069] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "3987d71f-47ab-4dd4-8426-95d78503effa" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1238.602319] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1238.790316] env[65522]: WARNING neutronclient.v2_0.client [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1238.802726] env[65522]: DEBUG oslo_vmware.api [None req-c9db8083-60e3-4aa4-abd3-418d2acbc584 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115085, 'name': SuspendVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.821376] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1238.821569] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1238.821773] env[65522]: DEBUG nova.network.neutron [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1238.931774] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115084, 'name': ReconfigVM_Task, 'duration_secs': 1.261711} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.932207] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1238.932493] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1de40a97-c0b8-4dde-bbd7-16731fa4f8f2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.939648] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1238.939648] env[65522]: value = "task-5115087" [ 1238.939648] env[65522]: _type = "Task" [ 1238.939648] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.948538] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115087, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.967251] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115086, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.108583] env[65522]: DEBUG nova.compute.utils [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1239.311192] env[65522]: DEBUG oslo_vmware.api [None req-c9db8083-60e3-4aa4-abd3-418d2acbc584 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115085, 'name': SuspendVM_Task} progress is 79%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.325505] env[65522]: WARNING neutronclient.v2_0.client [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1239.326767] env[65522]: WARNING openstack [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1239.328410] env[65522]: WARNING openstack [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1239.450811] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115087, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.467330] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115086, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.540666] env[65522]: WARNING openstack [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1239.541176] env[65522]: WARNING openstack [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1239.612318] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1239.689396] env[65522]: WARNING neutronclient.v2_0.client [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1239.690234] env[65522]: WARNING openstack [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1239.690821] env[65522]: WARNING openstack [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1239.805069] env[65522]: DEBUG oslo_vmware.api [None req-c9db8083-60e3-4aa4-abd3-418d2acbc584 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115085, 'name': SuspendVM_Task, 'duration_secs': 1.787002} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.805375] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c9db8083-60e3-4aa4-abd3-418d2acbc584 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Suspended the VM {{(pid=65522) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1239.805558] env[65522]: DEBUG nova.compute.manager [None req-c9db8083-60e3-4aa4-abd3-418d2acbc584 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1239.806373] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da84a41f-0b21-4add-91c1-2ff594bf8fb9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.811011] env[65522]: DEBUG nova.network.neutron [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating instance_info_cache with network_info: [{"id": "6b80135d-6228-4366-83aa-ff70808ae005", "address": "fa:16:3e:82:79:a0", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b80135d-62", "ovs_interfaceid": "6b80135d-6228-4366-83aa-ff70808ae005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1239.953207] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115087, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.966594] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115086, 'name': ReconfigVM_Task, 'duration_secs': 1.0974} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.966795] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 3b82f9ce-ea59-430d-adb6-918cc6cc48a4/3b82f9ce-ea59-430d-adb6-918cc6cc48a4.vmdk or device None with type streamOptimized {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1239.967481] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb786771-5147-4aa7-9800-f9da429a5a3b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.973607] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1239.973607] env[65522]: value = "task-5115088" [ 1239.973607] env[65522]: _type = "Task" [ 1239.973607] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.983753] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115088, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.313811] env[65522]: DEBUG oslo_concurrency.lockutils [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1240.454336] env[65522]: DEBUG oslo_vmware.api [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115087, 'name': PowerOnVM_Task, 'duration_secs': 1.033073} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.454673] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1240.457894] env[65522]: DEBUG nova.compute.manager [None req-b615127f-ec47-48e1-9359-749a5d39729b tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1240.458965] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7620fb56-ee98-4fe8-b1f6-02c349506f55 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.484334] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115088, 'name': Rename_Task, 'duration_secs': 0.420336} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.484612] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1240.484869] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7323c9f8-8edc-48d6-ac47-f5e09dbdd9b2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.492727] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1240.492727] env[65522]: value = "task-5115089" [ 1240.492727] env[65522]: _type = "Task" [ 1240.492727] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.504543] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115089, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.700253] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "3987d71f-47ab-4dd4-8426-95d78503effa" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1240.700678] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1240.700764] env[65522]: INFO nova.compute.manager [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Attaching volume 7dceaf16-7c29-4b7d-8595-65bb03d1aff5 to /dev/sdc [ 1240.738634] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca569706-cba0-4851-999b-a666a4b9554e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.748010] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337ef58e-864e-4f88-aeb7-4309c9c96096 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.762919] env[65522]: DEBUG nova.virt.block_device [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Updating existing volume attachment record: edb28d7c-ab12-468e-b0cd-891d09dc530d {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1240.824658] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae20959-e4e8-4c62-80ca-6f2ca5da64a4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.832958] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d97b87-1e5c-4c90-86ed-e65fbdf0f1f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.005214] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115089, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.233589] env[65522]: INFO nova.compute.manager [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Unrescuing [ 1241.233999] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "refresh_cache-874813b4-d0d6-4559-82e2-9fa2ea4feb15" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.234272] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquired lock "refresh_cache-874813b4-d0d6-4559-82e2-9fa2ea4feb15" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1241.234576] env[65522]: DEBUG nova.network.neutron [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1241.506132] env[65522]: DEBUG oslo_vmware.api [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115089, 'name': PowerOnVM_Task, 'duration_secs': 0.731157} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.506844] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1241.546929] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquiring lock "c1aca457-dc7c-477c-9b4f-3ed577fb2adb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1241.547191] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Lock "c1aca457-dc7c-477c-9b4f-3ed577fb2adb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1241.568209] env[65522]: INFO nova.compute.manager [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Resuming [ 1241.569288] env[65522]: DEBUG nova.objects.instance [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lazy-loading 'flavor' on Instance uuid 2981b635-43c8-4bd6-9991-e6af0be82f3c {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1241.626342] env[65522]: DEBUG nova.compute.manager [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1241.627201] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e528e8ef-6e4f-4069-9a93-de5995682f71 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.738323] env[65522]: WARNING neutronclient.v2_0.client [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1241.739222] env[65522]: WARNING openstack [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1241.739664] env[65522]: WARNING openstack [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1241.861547] env[65522]: WARNING openstack [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1241.862106] env[65522]: WARNING openstack [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1241.912883] env[65522]: WARNING neutronclient.v2_0.client [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1241.913661] env[65522]: WARNING openstack [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1241.914135] env[65522]: WARNING openstack [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1241.959554] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9d3c40-0749-488c-8ae6-a8decad26251 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.984410] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2af2ac01-00f8-464d-9620-68863418ce65 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.992490] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating instance 'daffdd40-9ee3-4a80-b670-5772395a32ae' progress to 83 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1242.002433] env[65522]: DEBUG nova.network.neutron [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Updating instance_info_cache with network_info: [{"id": "6ff7afa0-2b0e-450c-a755-77a3241a2527", "address": "fa:16:3e:c7:7a:9b", "network": {"id": "26dda233-0dd0-40ae-b56f-c84b1b91e92a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1197499239-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "15229af4da7a4ff9963369df50fb9d5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ff7afa0-2b", "ovs_interfaceid": "6ff7afa0-2b0e-450c-a755-77a3241a2527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1242.050228] env[65522]: DEBUG nova.compute.manager [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1242.144562] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a3e22380-1c80-4dfc-90f4-a12e882437ae tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 27.875s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1242.500069] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1242.500414] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7d30a166-5886-4cda-a03b-8eb2a0620e5a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.505372] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Releasing lock "refresh_cache-874813b4-d0d6-4559-82e2-9fa2ea4feb15" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1242.506038] env[65522]: DEBUG nova.objects.instance [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lazy-loading 'flavor' on Instance uuid 874813b4-d0d6-4559-82e2-9fa2ea4feb15 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1242.509128] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1242.509128] env[65522]: value = "task-5115091" [ 1242.509128] env[65522]: _type = "Task" [ 1242.509128] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.519222] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115091, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.573247] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1242.573546] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1242.575222] env[65522]: INFO nova.compute.claims [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1243.015994] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ffef0f-f566-48f0-b06e-46c235c43a4b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.024486] env[65522]: DEBUG oslo_vmware.api [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115091, 'name': PowerOnVM_Task, 'duration_secs': 0.426178} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.040439] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1243.040655] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-02b94b36-d6a0-4c6a-a98e-f1d3187ef391 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating instance 'daffdd40-9ee3-4a80-b670-5772395a32ae' progress to 100 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1243.044423] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1243.045000] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ed4b297-0e6b-4157-b1cb-ded30b51aa47 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.052009] env[65522]: DEBUG oslo_vmware.api [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1243.052009] env[65522]: value = "task-5115093" [ 1243.052009] env[65522]: _type = "Task" [ 1243.052009] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.060608] env[65522]: DEBUG oslo_vmware.api [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115093, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.083022] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.083199] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquired lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1243.084208] env[65522]: DEBUG nova.network.neutron [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1243.118025] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1243.118025] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1243.562405] env[65522]: DEBUG oslo_vmware.api [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115093, 'name': PowerOffVM_Task, 'duration_secs': 0.312508} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.562757] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1243.568124] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Reconfiguring VM instance instance-00000076 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1243.568423] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-287c7ecb-3d7e-4226-8502-cdf3acc07b79 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.588113] env[65522]: WARNING neutronclient.v2_0.client [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1243.588770] env[65522]: WARNING openstack [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1243.589143] env[65522]: WARNING openstack [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1243.600314] env[65522]: DEBUG oslo_vmware.api [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1243.600314] env[65522]: value = "task-5115094" [ 1243.600314] env[65522]: _type = "Task" [ 1243.600314] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.610948] env[65522]: DEBUG oslo_vmware.api [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115094, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.623989] env[65522]: DEBUG nova.compute.utils [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1243.710347] env[65522]: WARNING openstack [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1243.710729] env[65522]: WARNING openstack [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1243.778045] env[65522]: WARNING neutronclient.v2_0.client [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1243.778734] env[65522]: WARNING openstack [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1243.779106] env[65522]: WARNING openstack [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1243.788318] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6200c633-1fec-45a2-b6d2-a3bce1cf733a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.796740] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a16b04ea-bd94-4d64-93a1-9f34e0dc2751 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.833081] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12fcc1c9-c795-4109-95c3-c11818e70f99 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.841741] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1652ca9-3a95-43a8-a65d-9bb940a7cb19 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.855769] env[65522]: DEBUG nova.compute.provider_tree [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1243.875309] env[65522]: DEBUG nova.network.neutron [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Updating instance_info_cache with network_info: [{"id": "0c495486-a850-4368-95a9-03fac8060924", "address": "fa:16:3e:a7:35:38", "network": {"id": "2eb73cd0-541f-42e9-8774-600c5c1e9200", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-939435024-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fdba2dbc660145b7a98af8bd3f77a035", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c495486-a8", "ovs_interfaceid": "0c495486-a850-4368-95a9-03fac8060924", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1244.110833] env[65522]: DEBUG oslo_vmware.api [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115094, 'name': ReconfigVM_Task, 'duration_secs': 0.26526} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.111161] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Reconfigured VM instance instance-00000076 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1244.111372] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1244.111642] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fa442e42-96e6-4d8a-9adc-e3a66a3714c8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.119214] env[65522]: DEBUG oslo_vmware.api [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1244.119214] env[65522]: value = "task-5115095" [ 1244.119214] env[65522]: _type = "Task" [ 1244.119214] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.127628] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1244.133399] env[65522]: DEBUG oslo_vmware.api [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115095, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.359091] env[65522]: DEBUG nova.scheduler.client.report [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1244.378625] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Releasing lock "refresh_cache-2981b635-43c8-4bd6-9991-e6af0be82f3c" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1244.380172] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48690748-be59-4ac7-98e5-473900f43485 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.387365] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Resuming the VM {{(pid=65522) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1244.387643] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-adfdc430-5f49-4b79-99f0-76984f403753 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.395680] env[65522]: DEBUG oslo_vmware.api [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1244.395680] env[65522]: value = "task-5115096" [ 1244.395680] env[65522]: _type = "Task" [ 1244.395680] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.404157] env[65522]: DEBUG oslo_vmware.api [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115096, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.632611] env[65522]: DEBUG oslo_vmware.api [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115095, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.864789] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.291s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1244.865373] env[65522]: DEBUG nova.compute.manager [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1244.909158] env[65522]: DEBUG oslo_vmware.api [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115096, 'name': PowerOnVM_Task} progress is 90%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.068507] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "daffdd40-9ee3-4a80-b670-5772395a32ae" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1245.068987] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "daffdd40-9ee3-4a80-b670-5772395a32ae" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1245.069362] env[65522]: DEBUG nova.compute.manager [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Going to confirm migration 9 {{(pid=65522) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 1245.133211] env[65522]: DEBUG oslo_vmware.api [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115095, 'name': PowerOnVM_Task, 'duration_secs': 0.538958} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.133476] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1245.133626] env[65522]: DEBUG nova.compute.manager [None req-b08e69a4-fce4-453b-bfde-e821d22d4327 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1245.134617] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16205d5d-c397-4707-bd79-fff82322878f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.194469] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1245.194768] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1245.195146] env[65522]: INFO nova.compute.manager [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Attaching volume 25a65c4f-a464-4ece-b85f-95bc2dfd3903 to /dev/sdb [ 1245.232728] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-324e77d2-726d-4924-aacd-5b5bc48e7a00 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.244535] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98cbfbeb-469a-4175-86b0-145e766bf9e9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.265968] env[65522]: DEBUG nova.virt.block_device [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating existing volume attachment record: 94401581-9b5f-4a7d-83e7-7cc425560247 {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1245.327991] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1245.328332] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994976', 'volume_id': '7dceaf16-7c29-4b7d-8595-65bb03d1aff5', 'name': 'volume-7dceaf16-7c29-4b7d-8595-65bb03d1aff5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3987d71f-47ab-4dd4-8426-95d78503effa', 'attached_at': '', 'detached_at': '', 'volume_id': '7dceaf16-7c29-4b7d-8595-65bb03d1aff5', 'serial': '7dceaf16-7c29-4b7d-8595-65bb03d1aff5'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1245.329223] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20b315c-445a-45cc-8d7e-85ea8ae05c77 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.350049] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-addad38b-017d-4fee-a1eb-6afd085b4b6d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.375093] env[65522]: DEBUG nova.compute.utils [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1245.384705] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] volume-7dceaf16-7c29-4b7d-8595-65bb03d1aff5/volume-7dceaf16-7c29-4b7d-8595-65bb03d1aff5.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1245.385191] env[65522]: DEBUG nova.compute.manager [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Not allocating networking since 'none' was specified. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 1245.385437] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f93381a-d5fa-4784-b9c5-4ec8ff5d0619 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.411154] env[65522]: DEBUG oslo_vmware.api [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115096, 'name': PowerOnVM_Task, 'duration_secs': 0.76514} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.412553] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Resumed the VM {{(pid=65522) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1245.412738] env[65522]: DEBUG nova.compute.manager [None req-f60a4168-6d2e-4717-8d01-a5664f0fe931 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1245.413076] env[65522]: DEBUG oslo_vmware.api [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1245.413076] env[65522]: value = "task-5115097" [ 1245.413076] env[65522]: _type = "Task" [ 1245.413076] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.413784] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09630e53-22bc-4c7d-8d5c-c90e6bbb24ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.428979] env[65522]: DEBUG oslo_vmware.api [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.576280] env[65522]: WARNING neutronclient.v2_0.client [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1245.642125] env[65522]: WARNING neutronclient.v2_0.client [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1245.642476] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1245.642626] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquired lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1245.642809] env[65522]: DEBUG nova.network.neutron [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1245.642992] env[65522]: DEBUG nova.objects.instance [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lazy-loading 'info_cache' on Instance uuid daffdd40-9ee3-4a80-b670-5772395a32ae {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1245.886373] env[65522]: DEBUG nova.compute.manager [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1245.928522] env[65522]: DEBUG oslo_vmware.api [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115097, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.992905] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1245.993197] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1245.993414] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1245.993608] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1245.993783] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1245.996098] env[65522]: INFO nova.compute.manager [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Terminating instance [ 1246.426890] env[65522]: DEBUG oslo_vmware.api [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115097, 'name': ReconfigVM_Task, 'duration_secs': 0.836527} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.427823] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Reconfigured VM instance instance-00000070 to attach disk [datastore1] volume-7dceaf16-7c29-4b7d-8595-65bb03d1aff5/volume-7dceaf16-7c29-4b7d-8595-65bb03d1aff5.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1246.433869] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e77b740d-b262-48e1-b2da-29ea82d1fc7a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.452531] env[65522]: DEBUG oslo_vmware.api [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1246.452531] env[65522]: value = "task-5115101" [ 1246.452531] env[65522]: _type = "Task" [ 1246.452531] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.461605] env[65522]: DEBUG oslo_vmware.api [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115101, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.500311] env[65522]: DEBUG nova.compute.manager [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1246.500647] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1246.501971] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-590b7fa8-e0a4-4043-a522-306a5c4f27de {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.511161] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1246.511454] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-943e1bf1-6082-47d5-826c-bc779c7fa274 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.519884] env[65522]: DEBUG oslo_vmware.api [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1246.519884] env[65522]: value = "task-5115102" [ 1246.519884] env[65522]: _type = "Task" [ 1246.519884] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.533176] env[65522]: DEBUG oslo_vmware.api [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115102, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.650569] env[65522]: WARNING neutronclient.v2_0.client [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1246.651427] env[65522]: WARNING openstack [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1246.651804] env[65522]: WARNING openstack [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1246.776775] env[65522]: WARNING openstack [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1246.777188] env[65522]: WARNING openstack [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1246.857888] env[65522]: WARNING neutronclient.v2_0.client [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1246.858716] env[65522]: WARNING openstack [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1246.860052] env[65522]: WARNING openstack [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1246.897919] env[65522]: DEBUG nova.compute.manager [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1246.928034] env[65522]: DEBUG nova.virt.hardware [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1246.928302] env[65522]: DEBUG nova.virt.hardware [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1246.928456] env[65522]: DEBUG nova.virt.hardware [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1246.928634] env[65522]: DEBUG nova.virt.hardware [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1246.928777] env[65522]: DEBUG nova.virt.hardware [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1246.928923] env[65522]: DEBUG nova.virt.hardware [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1246.929151] env[65522]: DEBUG nova.virt.hardware [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1246.929314] env[65522]: DEBUG nova.virt.hardware [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1246.929479] env[65522]: DEBUG nova.virt.hardware [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1246.929641] env[65522]: DEBUG nova.virt.hardware [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1246.929813] env[65522]: DEBUG nova.virt.hardware [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1246.930794] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1300df16-1f1b-4033-9030-2d71df4118eb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.939390] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-025316f8-a252-471c-9944-35871b51c2e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.955134] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1246.961147] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Creating folder: Project (7b97c46719a84952862b04fa67eaf8c3). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1246.963691] env[65522]: DEBUG nova.network.neutron [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating instance_info_cache with network_info: [{"id": "6b80135d-6228-4366-83aa-ff70808ae005", "address": "fa:16:3e:82:79:a0", "network": {"id": "adc467d2-1ed1-4732-a996-8f7b83cb7b39", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-148035920-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "91597233ae9c44c094f4c32d90332fa6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b80135d-62", "ovs_interfaceid": "6b80135d-6228-4366-83aa-ff70808ae005", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1246.968653] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-571c415c-62cf-4bb3-837c-1f41e2e1ddfc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.977740] env[65522]: DEBUG oslo_vmware.api [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115101, 'name': ReconfigVM_Task, 'duration_secs': 0.197761} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.978088] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994976', 'volume_id': '7dceaf16-7c29-4b7d-8595-65bb03d1aff5', 'name': 'volume-7dceaf16-7c29-4b7d-8595-65bb03d1aff5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3987d71f-47ab-4dd4-8426-95d78503effa', 'attached_at': '', 'detached_at': '', 'volume_id': '7dceaf16-7c29-4b7d-8595-65bb03d1aff5', 'serial': '7dceaf16-7c29-4b7d-8595-65bb03d1aff5'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1246.982100] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Created folder: Project (7b97c46719a84952862b04fa67eaf8c3) in parent group-v994660. [ 1246.982317] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Creating folder: Instances. Parent ref: group-v994979. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1246.982581] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-da67991c-cb96-49a9-8ca8-1e47609ff533 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.994451] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Created folder: Instances in parent group-v994979. [ 1246.994524] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1246.995055] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1246.995327] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0bcf2ce4-0ac9-4fdd-bca0-a330341b6080 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.017316] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1247.017316] env[65522]: value = "task-5115105" [ 1247.017316] env[65522]: _type = "Task" [ 1247.017316] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.031554] env[65522]: DEBUG oslo_vmware.api [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115102, 'name': PowerOffVM_Task, 'duration_secs': 0.395602} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.035242] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1247.035434] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1247.035700] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115105, 'name': CreateVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.035932] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8386874c-c768-4ddb-9f05-ddaa433c4ae8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.115900] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1247.116230] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1247.116440] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Deleting the datastore file [datastore2] 874813b4-d0d6-4559-82e2-9fa2ea4feb15 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1247.116730] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d5db5c8d-e967-426c-a9fc-ebfc847265f4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.124458] env[65522]: DEBUG oslo_vmware.api [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1247.124458] env[65522]: value = "task-5115107" [ 1247.124458] env[65522]: _type = "Task" [ 1247.124458] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.134816] env[65522]: DEBUG oslo_vmware.api [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115107, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.467076] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Releasing lock "refresh_cache-daffdd40-9ee3-4a80-b670-5772395a32ae" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1247.467276] env[65522]: DEBUG nova.objects.instance [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lazy-loading 'migration_context' on Instance uuid daffdd40-9ee3-4a80-b670-5772395a32ae {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1247.528098] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115105, 'name': CreateVM_Task, 'duration_secs': 0.380057} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.528240] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1247.528642] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1247.528794] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1247.529134] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1247.529420] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6428ad54-7ea9-4c5e-b194-bd2f6f63c12f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.535404] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1247.535404] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529a4b6b-40fc-8876-390b-ebc04c598915" [ 1247.535404] env[65522]: _type = "Task" [ 1247.535404] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.544840] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529a4b6b-40fc-8876-390b-ebc04c598915, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.635650] env[65522]: DEBUG oslo_vmware.api [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115107, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157462} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.635919] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1247.636129] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1247.636311] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1247.636482] env[65522]: INFO nova.compute.manager [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1247.636731] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1247.636927] env[65522]: DEBUG nova.compute.manager [-] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1247.637037] env[65522]: DEBUG nova.network.neutron [-] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1247.637284] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1247.637812] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1247.638082] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1247.687685] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1247.951433] env[65522]: DEBUG nova.compute.manager [req-3d36e128-8b58-4911-9ec9-17827ea4daa6 req-f3ccbeca-3698-4733-b7b8-e210565b1215 service nova] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Received event network-vif-deleted-6ff7afa0-2b0e-450c-a755-77a3241a2527 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1247.952594] env[65522]: INFO nova.compute.manager [req-3d36e128-8b58-4911-9ec9-17827ea4daa6 req-f3ccbeca-3698-4733-b7b8-e210565b1215 service nova] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Neutron deleted interface 6ff7afa0-2b0e-450c-a755-77a3241a2527; detaching it from the instance and deleting it from the info cache [ 1247.952594] env[65522]: DEBUG nova.network.neutron [req-3d36e128-8b58-4911-9ec9-17827ea4daa6 req-f3ccbeca-3698-4733-b7b8-e210565b1215 service nova] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1247.970447] env[65522]: DEBUG nova.objects.base [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1247.971663] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bfd44c-bae9-4ec0-a8fe-b1cc83334b67 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.992474] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7ddd02a-8cda-4350-a29c-898997a65725 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.998654] env[65522]: DEBUG oslo_vmware.api [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1247.998654] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5278f3ae-7fa6-4bbd-e2e6-bef6615e51d8" [ 1247.998654] env[65522]: _type = "Task" [ 1247.998654] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.008637] env[65522]: DEBUG oslo_vmware.api [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5278f3ae-7fa6-4bbd-e2e6-bef6615e51d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.016015] env[65522]: DEBUG nova.objects.instance [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lazy-loading 'flavor' on Instance uuid 3987d71f-47ab-4dd4-8426-95d78503effa {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1248.046612] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529a4b6b-40fc-8876-390b-ebc04c598915, 'name': SearchDatastore_Task, 'duration_secs': 0.01319} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.046948] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1248.047249] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1248.047503] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1248.047651] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1248.047853] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1248.048151] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a3114e65-c35d-4ad5-93a4-aa054c14ef88 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.058373] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1248.058512] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1248.059325] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9d9fc2c-9bea-4219-a9c1-22694af4a710 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.065848] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1248.065848] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529c3bbc-7cca-9d08-8456-7b1de4ce4185" [ 1248.065848] env[65522]: _type = "Task" [ 1248.065848] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.077487] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529c3bbc-7cca-9d08-8456-7b1de4ce4185, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.389662] env[65522]: DEBUG nova.network.neutron [-] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1248.454740] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b458dcc9-3c8a-496e-bfb3-a5bb2c9a4f1c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.464863] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eca19f1-8f8d-441a-a27b-7d677acad4a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.476192] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "3987d71f-47ab-4dd4-8426-95d78503effa" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1248.501716] env[65522]: DEBUG nova.compute.manager [req-3d36e128-8b58-4911-9ec9-17827ea4daa6 req-f3ccbeca-3698-4733-b7b8-e210565b1215 service nova] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Detach interface failed, port_id=6ff7afa0-2b0e-450c-a755-77a3241a2527, reason: Instance 874813b4-d0d6-4559-82e2-9fa2ea4feb15 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1248.510837] env[65522]: DEBUG oslo_vmware.api [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5278f3ae-7fa6-4bbd-e2e6-bef6615e51d8, 'name': SearchDatastore_Task, 'duration_secs': 0.01546} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.511149] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1248.511529] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1248.520272] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b5124434-8107-4460-b8eb-c73989d3a113 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.820s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1248.521021] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.045s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1248.577120] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529c3bbc-7cca-9d08-8456-7b1de4ce4185, 'name': SearchDatastore_Task, 'duration_secs': 0.011063} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.578062] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e7b174d-96b3-4044-846c-6a01b8984e41 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.583694] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1248.583694] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c5378f-0dca-ea6b-69c3-077a23bfe185" [ 1248.583694] env[65522]: _type = "Task" [ 1248.583694] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.592010] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c5378f-0dca-ea6b-69c3-077a23bfe185, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.893543] env[65522]: INFO nova.compute.manager [-] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Took 1.26 seconds to deallocate network for instance. [ 1249.023594] env[65522]: INFO nova.compute.manager [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Detaching volume 43d813d9-8062-43f5-8ca7-9f86fdd62d1c [ 1249.054019] env[65522]: INFO nova.virt.block_device [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Attempting to driver detach volume 43d813d9-8062-43f5-8ca7-9f86fdd62d1c from mountpoint /dev/sdb [ 1249.054297] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1249.054511] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994975', 'volume_id': '43d813d9-8062-43f5-8ca7-9f86fdd62d1c', 'name': 'volume-43d813d9-8062-43f5-8ca7-9f86fdd62d1c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3987d71f-47ab-4dd4-8426-95d78503effa', 'attached_at': '', 'detached_at': '', 'volume_id': '43d813d9-8062-43f5-8ca7-9f86fdd62d1c', 'serial': '43d813d9-8062-43f5-8ca7-9f86fdd62d1c'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1249.055497] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da93321a-f240-49d6-a2cc-dca4364a86f9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.084144] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a660d899-ac5c-4c10-ae74-77354c20f782 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.096255] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c5378f-0dca-ea6b-69c3-077a23bfe185, 'name': SearchDatastore_Task, 'duration_secs': 0.01157} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.100666] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1249.100939] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] c1aca457-dc7c-477c-9b4f-3ed577fb2adb/c1aca457-dc7c-477c-9b4f-3ed577fb2adb.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1249.101618] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-317df11c-4e09-4405-a17b-a41b5284dea3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.104132] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67241bfa-d3ca-4e4c-b880-9d006b0f940d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.133704] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bdc214d-54b0-42f6-b471-f553712c1e7f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.136359] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1249.136359] env[65522]: value = "task-5115109" [ 1249.136359] env[65522]: _type = "Task" [ 1249.136359] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.150480] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The volume has not been displaced from its original location: [datastore1] volume-43d813d9-8062-43f5-8ca7-9f86fdd62d1c/volume-43d813d9-8062-43f5-8ca7-9f86fdd62d1c.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1249.155926] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Reconfiguring VM instance instance-00000070 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1249.158998] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6683a63-9950-41e4-89e3-22d29821e6f5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.175468] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115109, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.181476] env[65522]: DEBUG oslo_vmware.api [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1249.181476] env[65522]: value = "task-5115110" [ 1249.181476] env[65522]: _type = "Task" [ 1249.181476] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.191108] env[65522]: DEBUG oslo_vmware.api [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115110, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.272176] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd406d5-781a-4ad6-b3d2-ae4f104ee8df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.281196] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ef48f9-2b22-47a9-ba25-a20cf3c73550 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.320306] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e764a4c8-6ba9-4091-9482-9d0861a41a7d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.331748] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb5eef02-3d73-44be-ba34-f0d31f14cbee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.348780] env[65522]: DEBUG nova.compute.provider_tree [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1249.400966] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1249.649332] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115109, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476686} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.649637] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] c1aca457-dc7c-477c-9b4f-3ed577fb2adb/c1aca457-dc7c-477c-9b4f-3ed577fb2adb.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1249.649942] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1249.650284] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8c0033e-046e-41ab-b19b-f8e1117e33d8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.661059] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1249.661059] env[65522]: value = "task-5115111" [ 1249.661059] env[65522]: _type = "Task" [ 1249.661059] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.670648] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115111, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.692921] env[65522]: DEBUG oslo_vmware.api [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115110, 'name': ReconfigVM_Task, 'duration_secs': 0.314526} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.693300] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Reconfigured VM instance instance-00000070 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1249.699191] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d0755e1-ff83-43c7-8db9-453655bfee60 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.716329] env[65522]: DEBUG oslo_vmware.api [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1249.716329] env[65522]: value = "task-5115112" [ 1249.716329] env[65522]: _type = "Task" [ 1249.716329] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.726162] env[65522]: DEBUG oslo_vmware.api [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115112, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.852995] env[65522]: DEBUG nova.scheduler.client.report [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1250.104443] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1250.104988] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1250.105135] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "2981b635-43c8-4bd6-9991-e6af0be82f3c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1250.105393] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1250.105579] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1250.107957] env[65522]: INFO nova.compute.manager [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Terminating instance [ 1250.173079] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115111, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078391} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.173408] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1250.174667] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09fe0bdc-c16f-46a5-9799-1cfdb74c9c29 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.202452] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Reconfiguring VM instance instance-00000077 to attach disk [datastore1] c1aca457-dc7c-477c-9b4f-3ed577fb2adb/c1aca457-dc7c-477c-9b4f-3ed577fb2adb.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1250.202834] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb05e9c8-98d7-46ee-99fd-c4840df64149 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.227916] env[65522]: DEBUG oslo_vmware.api [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115112, 'name': ReconfigVM_Task, 'duration_secs': 0.162763} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.231037] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994975', 'volume_id': '43d813d9-8062-43f5-8ca7-9f86fdd62d1c', 'name': 'volume-43d813d9-8062-43f5-8ca7-9f86fdd62d1c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3987d71f-47ab-4dd4-8426-95d78503effa', 'attached_at': '', 'detached_at': '', 'volume_id': '43d813d9-8062-43f5-8ca7-9f86fdd62d1c', 'serial': '43d813d9-8062-43f5-8ca7-9f86fdd62d1c'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1250.234061] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1250.234061] env[65522]: value = "task-5115113" [ 1250.234061] env[65522]: _type = "Task" [ 1250.234061] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.253229] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115113, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.324776] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1250.324776] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994978', 'volume_id': '25a65c4f-a464-4ece-b85f-95bc2dfd3903', 'name': 'volume-25a65c4f-a464-4ece-b85f-95bc2dfd3903', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3adfefa5-fb19-44b2-b3c5-42f2e2918673', 'attached_at': '', 'detached_at': '', 'volume_id': '25a65c4f-a464-4ece-b85f-95bc2dfd3903', 'serial': '25a65c4f-a464-4ece-b85f-95bc2dfd3903'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1250.325736] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af8eedd-0bc4-4452-bf50-def31b3bd45f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.343765] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c7bffc-11bf-45ef-9970-864d70036495 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.373994] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] volume-25a65c4f-a464-4ece-b85f-95bc2dfd3903/volume-25a65c4f-a464-4ece-b85f-95bc2dfd3903.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1250.374466] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-210b6eee-522b-4461-b919-41325c27a08e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.395546] env[65522]: DEBUG oslo_vmware.api [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1250.395546] env[65522]: value = "task-5115114" [ 1250.395546] env[65522]: _type = "Task" [ 1250.395546] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.405640] env[65522]: DEBUG oslo_vmware.api [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115114, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.612057] env[65522]: DEBUG nova.compute.manager [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1250.612241] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1250.613138] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-922aa5ae-742c-460f-862f-1fbe6503035f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.622609] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1250.622947] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d3676301-5c16-4816-bfcd-d55c8bff8fe2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.630238] env[65522]: DEBUG oslo_vmware.api [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1250.630238] env[65522]: value = "task-5115115" [ 1250.630238] env[65522]: _type = "Task" [ 1250.630238] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.640627] env[65522]: DEBUG oslo_vmware.api [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115115, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.746605] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115113, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.781454] env[65522]: DEBUG nova.objects.instance [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lazy-loading 'flavor' on Instance uuid 3987d71f-47ab-4dd4-8426-95d78503effa {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1250.876729] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.365s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1250.881181] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.480s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1250.881592] env[65522]: DEBUG nova.objects.instance [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lazy-loading 'resources' on Instance uuid 874813b4-d0d6-4559-82e2-9fa2ea4feb15 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1250.906225] env[65522]: DEBUG oslo_vmware.api [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115114, 'name': ReconfigVM_Task, 'duration_secs': 0.508118} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.906667] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Reconfigured VM instance instance-00000073 to attach disk [datastore1] volume-25a65c4f-a464-4ece-b85f-95bc2dfd3903/volume-25a65c4f-a464-4ece-b85f-95bc2dfd3903.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1250.911667] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b604a07-0d91-4747-884b-5aec187b3e76 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.927943] env[65522]: DEBUG oslo_vmware.api [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1250.927943] env[65522]: value = "task-5115116" [ 1250.927943] env[65522]: _type = "Task" [ 1250.927943] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.937467] env[65522]: DEBUG oslo_vmware.api [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115116, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.141114] env[65522]: DEBUG oslo_vmware.api [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115115, 'name': PowerOffVM_Task, 'duration_secs': 0.292618} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.141432] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1251.141565] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1251.141853] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4575b19-5553-4133-87c6-8db9fbe13ae6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.245081] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1251.245381] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1251.245640] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Deleting the datastore file [datastore2] 2981b635-43c8-4bd6-9991-e6af0be82f3c {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1251.246493] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e53c674f-af2e-457c-ba6c-4aa1bdab64bf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.256191] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115113, 'name': ReconfigVM_Task, 'duration_secs': 0.672943} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.257520] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Reconfigured VM instance instance-00000077 to attach disk [datastore1] c1aca457-dc7c-477c-9b4f-3ed577fb2adb/c1aca457-dc7c-477c-9b4f-3ed577fb2adb.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1251.258323] env[65522]: DEBUG oslo_vmware.api [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for the task: (returnval){ [ 1251.258323] env[65522]: value = "task-5115118" [ 1251.258323] env[65522]: _type = "Task" [ 1251.258323] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.258538] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-252c3dd9-7d49-47b5-bde5-ea78501acdc5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.269312] env[65522]: DEBUG oslo_vmware.api [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115118, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.271033] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1251.271033] env[65522]: value = "task-5115119" [ 1251.271033] env[65522]: _type = "Task" [ 1251.271033] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.442225] env[65522]: DEBUG oslo_vmware.api [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115116, 'name': ReconfigVM_Task, 'duration_secs': 0.236718} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.443081] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994978', 'volume_id': '25a65c4f-a464-4ece-b85f-95bc2dfd3903', 'name': 'volume-25a65c4f-a464-4ece-b85f-95bc2dfd3903', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3adfefa5-fb19-44b2-b3c5-42f2e2918673', 'attached_at': '', 'detached_at': '', 'volume_id': '25a65c4f-a464-4ece-b85f-95bc2dfd3903', 'serial': '25a65c4f-a464-4ece-b85f-95bc2dfd3903'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1251.445796] env[65522]: INFO nova.scheduler.client.report [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleted allocation for migration 5817abdf-1156-4dda-8607-b249c325820b [ 1251.576235] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90adf97a-094e-4020-8cf5-9f3f2fa147e8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.585241] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79fc6779-3d7d-4e69-989f-66f51bee7d5c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.619574] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0b16b7-d9df-4e6d-9031-af7af955edf2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.623452] env[65522]: INFO nova.compute.manager [None req-af33cc08-91f8-4ae3-bcac-33ab7a6a7284 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Get console output [ 1251.623797] env[65522]: WARNING nova.virt.vmwareapi.driver [None req-af33cc08-91f8-4ae3-bcac-33ab7a6a7284 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] The console log is missing. Check your VSPC configuration [ 1251.632904] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc22cb01-f2f6-40fd-b21d-8d20994e9aee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.648877] env[65522]: DEBUG nova.compute.provider_tree [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1251.742962] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "3987d71f-47ab-4dd4-8426-95d78503effa" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1251.770612] env[65522]: DEBUG oslo_vmware.api [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Task: {'id': task-5115118, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.301093} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.770858] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1251.771057] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1251.771237] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1251.771405] env[65522]: INFO nova.compute.manager [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1251.771646] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1251.771918] env[65522]: DEBUG nova.compute.manager [-] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1251.772022] env[65522]: DEBUG nova.network.neutron [-] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1251.772267] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1251.772791] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1251.773171] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1251.788892] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e66547d8-8190-49fd-bea1-403aac4d6a21 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.268s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1251.790067] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115119, 'name': Rename_Task, 'duration_secs': 0.276208} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.790260] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.048s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1251.791453] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1251.792229] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd47dcb4-0fbc-48c6-bb70-48484b146e98 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.802118] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1251.802118] env[65522]: value = "task-5115120" [ 1251.802118] env[65522]: _type = "Task" [ 1251.802118] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.814794] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115120, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.859726] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1251.953757] env[65522]: DEBUG oslo_concurrency.lockutils [None req-4e71888d-38b9-42cb-a00d-b6be5bdaa1a4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "daffdd40-9ee3-4a80-b670-5772395a32ae" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.885s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1252.119542] env[65522]: DEBUG nova.compute.manager [req-0bb9ad3f-6b33-4f2e-bcd5-598e83ceab2c req-cc1d13d8-4909-4b8b-95ee-9a937fe6c03f service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Received event network-vif-deleted-0c495486-a850-4368-95a9-03fac8060924 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1252.119622] env[65522]: INFO nova.compute.manager [req-0bb9ad3f-6b33-4f2e-bcd5-598e83ceab2c req-cc1d13d8-4909-4b8b-95ee-9a937fe6c03f service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Neutron deleted interface 0c495486-a850-4368-95a9-03fac8060924; detaching it from the instance and deleting it from the info cache [ 1252.119788] env[65522]: DEBUG nova.network.neutron [req-0bb9ad3f-6b33-4f2e-bcd5-598e83ceab2c req-cc1d13d8-4909-4b8b-95ee-9a937fe6c03f service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1252.152449] env[65522]: DEBUG nova.scheduler.client.report [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1252.293429] env[65522]: INFO nova.compute.manager [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Detaching volume 7dceaf16-7c29-4b7d-8595-65bb03d1aff5 [ 1252.312430] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115120, 'name': PowerOnVM_Task} progress is 87%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.325732] env[65522]: INFO nova.virt.block_device [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Attempting to driver detach volume 7dceaf16-7c29-4b7d-8595-65bb03d1aff5 from mountpoint /dev/sdc [ 1252.325978] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1252.326193] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994976', 'volume_id': '7dceaf16-7c29-4b7d-8595-65bb03d1aff5', 'name': 'volume-7dceaf16-7c29-4b7d-8595-65bb03d1aff5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3987d71f-47ab-4dd4-8426-95d78503effa', 'attached_at': '', 'detached_at': '', 'volume_id': '7dceaf16-7c29-4b7d-8595-65bb03d1aff5', 'serial': '7dceaf16-7c29-4b7d-8595-65bb03d1aff5'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1252.327223] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ade43e-b3a6-4667-9cf0-6cf8e745ea84 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.351262] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ab0b17-d9ef-41d9-8d88-56b9026030d5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.359170] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42a461f-8e8e-459c-bc55-bfda57315ba4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.400912] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6f5062-206f-4437-a9ba-997ea51f7243 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.425226] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] The volume has not been displaced from its original location: [datastore1] volume-7dceaf16-7c29-4b7d-8595-65bb03d1aff5/volume-7dceaf16-7c29-4b7d-8595-65bb03d1aff5.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1252.431010] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Reconfiguring VM instance instance-00000070 to detach disk 2002 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1252.431919] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9dfb34ad-fdb5-4354-978f-30f2ca751596 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.452942] env[65522]: DEBUG oslo_vmware.api [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1252.452942] env[65522]: value = "task-5115121" [ 1252.452942] env[65522]: _type = "Task" [ 1252.452942] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.464113] env[65522]: DEBUG oslo_vmware.api [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115121, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.491047] env[65522]: DEBUG nova.objects.instance [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'flavor' on Instance uuid 3adfefa5-fb19-44b2-b3c5-42f2e2918673 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1252.563142] env[65522]: DEBUG nova.network.neutron [-] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1252.623085] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3d890b37-a928-4ceb-abf0-144805e0f8b4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.633891] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6144dd87-b5c6-4196-a8cd-32fdaf67c6a0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.657525] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.776s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1252.671942] env[65522]: DEBUG nova.compute.manager [req-0bb9ad3f-6b33-4f2e-bcd5-598e83ceab2c req-cc1d13d8-4909-4b8b-95ee-9a937fe6c03f service nova] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Detach interface failed, port_id=0c495486-a850-4368-95a9-03fac8060924, reason: Instance 2981b635-43c8-4bd6-9991-e6af0be82f3c could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1252.685673] env[65522]: INFO nova.scheduler.client.report [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Deleted allocations for instance 874813b4-d0d6-4559-82e2-9fa2ea4feb15 [ 1252.815311] env[65522]: DEBUG oslo_vmware.api [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115120, 'name': PowerOnVM_Task, 'duration_secs': 0.735176} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.815558] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1252.815755] env[65522]: INFO nova.compute.manager [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Took 5.92 seconds to spawn the instance on the hypervisor. [ 1252.815932] env[65522]: DEBUG nova.compute.manager [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1252.816713] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f019e0cc-1e4a-4bdc-928c-120ff27c2bfc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.964113] env[65522]: DEBUG oslo_vmware.api [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115121, 'name': ReconfigVM_Task, 'duration_secs': 0.324106} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.964420] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Reconfigured VM instance instance-00000070 to detach disk 2002 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1252.969675] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-426ba86c-6dd4-4d03-a0de-05b634df3977 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.986364] env[65522]: DEBUG oslo_vmware.api [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1252.986364] env[65522]: value = "task-5115122" [ 1252.986364] env[65522]: _type = "Task" [ 1252.986364] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.997761] env[65522]: DEBUG oslo_vmware.api [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115122, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.999060] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f89e723a-654e-4334-a9de-4ee8c0965236 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.804s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1253.068110] env[65522]: INFO nova.compute.manager [-] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Took 1.29 seconds to deallocate network for instance. [ 1253.194694] env[65522]: DEBUG oslo_concurrency.lockutils [None req-19e48cd4-6ed4-409e-8d85-a4758f293319 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "874813b4-d0d6-4559-82e2-9fa2ea4feb15" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.201s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1253.332724] env[65522]: INFO nova.compute.manager [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Took 10.78 seconds to build instance. [ 1253.500265] env[65522]: DEBUG oslo_vmware.api [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115122, 'name': ReconfigVM_Task, 'duration_secs': 0.170551} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.500751] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994976', 'volume_id': '7dceaf16-7c29-4b7d-8595-65bb03d1aff5', 'name': 'volume-7dceaf16-7c29-4b7d-8595-65bb03d1aff5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3987d71f-47ab-4dd4-8426-95d78503effa', 'attached_at': '', 'detached_at': '', 'volume_id': '7dceaf16-7c29-4b7d-8595-65bb03d1aff5', 'serial': '7dceaf16-7c29-4b7d-8595-65bb03d1aff5'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1253.574911] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1253.574911] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1253.575220] env[65522]: DEBUG nova.objects.instance [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lazy-loading 'resources' on Instance uuid 2981b635-43c8-4bd6-9991-e6af0be82f3c {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1253.785086] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "874b6a07-ebba-4f75-937e-2281af1598b8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1253.785389] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "874b6a07-ebba-4f75-937e-2281af1598b8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1253.786075] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "874b6a07-ebba-4f75-937e-2281af1598b8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1253.786075] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "874b6a07-ebba-4f75-937e-2281af1598b8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1253.786075] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "874b6a07-ebba-4f75-937e-2281af1598b8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1253.788223] env[65522]: INFO nova.compute.manager [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Terminating instance [ 1253.834540] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0e22160f-3a08-4815-9b5a-65b600afacbd tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Lock "c1aca457-dc7c-477c-9b4f-3ed577fb2adb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.287s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1253.948530] env[65522]: INFO nova.compute.manager [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Rebuilding instance [ 1253.997951] env[65522]: DEBUG nova.compute.manager [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1253.999250] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab0b47a-edc6-4307-8772-aca75e6b6a62 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.051468] env[65522]: DEBUG nova.objects.instance [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lazy-loading 'flavor' on Instance uuid 3987d71f-47ab-4dd4-8426-95d78503effa {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1254.211557] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e7221d-bc76-4379-8db9-30075b53bb1e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.219600] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c3cb1c-248d-447b-a4a5-7cf2bea6effe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.249626] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73cde11f-0a2f-44ce-9f1d-f1946feb1cdf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.257545] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54fbb913-4805-4a48-b123-4af4028d9e4e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.270603] env[65522]: DEBUG nova.compute.provider_tree [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1254.292123] env[65522]: DEBUG nova.compute.manager [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1254.292360] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1254.293289] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99545c94-0607-45eb-969c-25976d8d0b87 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.301987] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1254.302230] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-30f6c02e-d99c-434d-9042-f6fd92d18eed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.308612] env[65522]: DEBUG oslo_vmware.api [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1254.308612] env[65522]: value = "task-5115123" [ 1254.308612] env[65522]: _type = "Task" [ 1254.308612] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.319065] env[65522]: DEBUG oslo_vmware.api [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115123, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.471068] env[65522]: DEBUG nova.compute.manager [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Stashing vm_state: active {{(pid=65522) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 1254.773598] env[65522]: DEBUG nova.scheduler.client.report [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1254.818560] env[65522]: DEBUG oslo_vmware.api [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115123, 'name': PowerOffVM_Task, 'duration_secs': 0.253112} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.818798] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1254.818971] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1254.819279] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b4f3e5fb-4bda-4c8f-97c1-7f0a7e20daae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.888329] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1254.888549] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1254.888734] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Deleting the datastore file [datastore2] 874b6a07-ebba-4f75-937e-2281af1598b8 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1254.889456] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1458d1b-2bf3-448c-8ebc-fe2e43b0024e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.898416] env[65522]: DEBUG oslo_vmware.api [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for the task: (returnval){ [ 1254.898416] env[65522]: value = "task-5115125" [ 1254.898416] env[65522]: _type = "Task" [ 1254.898416] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.907334] env[65522]: DEBUG oslo_vmware.api [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115125, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.992877] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1255.013821] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1255.014238] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-541f8e36-bdfa-4b3b-8955-0c86a09080d2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.022898] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1255.022898] env[65522]: value = "task-5115126" [ 1255.022898] env[65522]: _type = "Task" [ 1255.022898] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.033331] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115126, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.062199] env[65522]: DEBUG oslo_concurrency.lockutils [None req-bc6bd5f2-7244-4626-8b3a-1f3a66955556 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.272s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1255.278959] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.704s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1255.281395] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.289s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1255.301639] env[65522]: INFO nova.scheduler.client.report [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Deleted allocations for instance 2981b635-43c8-4bd6-9991-e6af0be82f3c [ 1255.325666] env[65522]: DEBUG oslo_concurrency.lockutils [None req-064ec591-c140-4a13-bfc2-9bec9f2e78a1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1255.325942] env[65522]: DEBUG oslo_concurrency.lockutils [None req-064ec591-c140-4a13-bfc2-9bec9f2e78a1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1255.326154] env[65522]: DEBUG nova.compute.manager [None req-064ec591-c140-4a13-bfc2-9bec9f2e78a1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1255.327263] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54eb560b-b270-48f2-958c-84f5225eea99 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.336135] env[65522]: DEBUG nova.compute.manager [None req-064ec591-c140-4a13-bfc2-9bec9f2e78a1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65522) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 1255.336776] env[65522]: DEBUG nova.objects.instance [None req-064ec591-c140-4a13-bfc2-9bec9f2e78a1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lazy-loading 'flavor' on Instance uuid 1b0fcdbb-aa9e-4915-a8c6-1167aa888390 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1255.410156] env[65522]: DEBUG oslo_vmware.api [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Task: {'id': task-5115125, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191857} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.410389] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1255.410579] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1255.410761] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1255.410938] env[65522]: INFO nova.compute.manager [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1255.411230] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1255.411444] env[65522]: DEBUG nova.compute.manager [-] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1255.411546] env[65522]: DEBUG nova.network.neutron [-] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1255.411801] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1255.412510] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1255.412782] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1255.451126] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1255.536593] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115126, 'name': PowerOffVM_Task, 'duration_secs': 0.210149} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.536729] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1255.538375] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1255.538527] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa7b2097-2787-4568-88cb-84e4ab6ed10a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.548138] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1255.548569] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b47c21e8-a454-4ad9-9a66-8bf05a1c0b60 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.577567] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1255.577793] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1255.577981] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Deleting the datastore file [datastore1] c1aca457-dc7c-477c-9b4f-3ed577fb2adb {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1255.578496] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1966c617-8127-42bd-ad7b-de5b887a5648 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.587203] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1255.587203] env[65522]: value = "task-5115128" [ 1255.587203] env[65522]: _type = "Task" [ 1255.587203] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.602341] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115128, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.741231] env[65522]: DEBUG nova.compute.manager [req-d0c3a416-0137-4d8d-bc32-0c0da1fd72ac req-701b75ae-d8cf-41b8-8387-eaa6635489cc service nova] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Received event network-vif-deleted-54d3b471-cec2-4c58-a51d-7bd3ed42e50e {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1255.741436] env[65522]: INFO nova.compute.manager [req-d0c3a416-0137-4d8d-bc32-0c0da1fd72ac req-701b75ae-d8cf-41b8-8387-eaa6635489cc service nova] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Neutron deleted interface 54d3b471-cec2-4c58-a51d-7bd3ed42e50e; detaching it from the instance and deleting it from the info cache [ 1255.741674] env[65522]: DEBUG nova.network.neutron [req-d0c3a416-0137-4d8d-bc32-0c0da1fd72ac req-701b75ae-d8cf-41b8-8387-eaa6635489cc service nova] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1255.789286] env[65522]: INFO nova.compute.claims [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1255.809474] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8e3c674d-b88e-41d1-94f7-661faa569f98 tempest-ServersNegativeTestJSON-711616081 tempest-ServersNegativeTestJSON-711616081-project-member] Lock "2981b635-43c8-4bd6-9991-e6af0be82f3c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.705s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1256.097736] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115128, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099497} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.098021] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1256.098205] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1256.098430] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1256.168445] env[65522]: DEBUG nova.network.neutron [-] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1256.209069] env[65522]: DEBUG oslo_concurrency.lockutils [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "3987d71f-47ab-4dd4-8426-95d78503effa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1256.209069] env[65522]: DEBUG oslo_concurrency.lockutils [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1256.209069] env[65522]: DEBUG oslo_concurrency.lockutils [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "3987d71f-47ab-4dd4-8426-95d78503effa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1256.209343] env[65522]: DEBUG oslo_concurrency.lockutils [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1256.209465] env[65522]: DEBUG oslo_concurrency.lockutils [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1256.211698] env[65522]: INFO nova.compute.manager [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Terminating instance [ 1256.243990] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3580e062-493a-4e1a-b12d-94041790e3fb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.257183] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed4c2785-02ab-4d77-aa6c-9aa64c8487b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.299548] env[65522]: INFO nova.compute.resource_tracker [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating resource usage from migration 9498bd7f-755a-48dd-81bd-f3fcf24eea86 [ 1256.302228] env[65522]: DEBUG nova.compute.manager [req-d0c3a416-0137-4d8d-bc32-0c0da1fd72ac req-701b75ae-d8cf-41b8-8387-eaa6635489cc service nova] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Detach interface failed, port_id=54d3b471-cec2-4c58-a51d-7bd3ed42e50e, reason: Instance 874b6a07-ebba-4f75-937e-2281af1598b8 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1256.344486] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-064ec591-c140-4a13-bfc2-9bec9f2e78a1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1256.345162] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-45c9dc4e-2613-4145-8e47-570286727144 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.352477] env[65522]: DEBUG oslo_vmware.api [None req-064ec591-c140-4a13-bfc2-9bec9f2e78a1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1256.352477] env[65522]: value = "task-5115129" [ 1256.352477] env[65522]: _type = "Task" [ 1256.352477] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.365980] env[65522]: DEBUG oslo_vmware.api [None req-064ec591-c140-4a13-bfc2-9bec9f2e78a1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115129, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.443283] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-366f679d-2ce4-4e57-a6bd-d5f43ff8ffe7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.451528] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc09dfd-829f-479d-b2e2-3fda03dcdcab {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.484620] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7685df-b5b1-4889-9677-a2fd005a4762 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.493106] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e2f4e6d-9169-47fd-a7c5-0d0ff41e6fd4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.508974] env[65522]: DEBUG nova.compute.provider_tree [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1256.671141] env[65522]: INFO nova.compute.manager [-] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Took 1.26 seconds to deallocate network for instance. [ 1256.715753] env[65522]: DEBUG nova.compute.manager [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1256.715988] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1256.716884] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728ba878-556b-4a6f-9c29-e8b0b4e432ff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.725770] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1256.726064] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c3f135a-419b-4cec-afda-a9761ffe6b2d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.733869] env[65522]: DEBUG oslo_vmware.api [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1256.733869] env[65522]: value = "task-5115130" [ 1256.733869] env[65522]: _type = "Task" [ 1256.733869] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.743546] env[65522]: DEBUG oslo_vmware.api [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115130, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.866242] env[65522]: DEBUG oslo_vmware.api [None req-064ec591-c140-4a13-bfc2-9bec9f2e78a1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115129, 'name': PowerOffVM_Task, 'duration_secs': 0.254903} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.866537] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-064ec591-c140-4a13-bfc2-9bec9f2e78a1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1256.866734] env[65522]: DEBUG nova.compute.manager [None req-064ec591-c140-4a13-bfc2-9bec9f2e78a1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1256.867693] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd92fdf-8b06-4489-912b-fe3970bb206a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.013129] env[65522]: DEBUG nova.scheduler.client.report [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1257.141813] env[65522]: DEBUG nova.virt.hardware [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1257.142347] env[65522]: DEBUG nova.virt.hardware [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1257.142534] env[65522]: DEBUG nova.virt.hardware [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1257.142719] env[65522]: DEBUG nova.virt.hardware [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1257.142862] env[65522]: DEBUG nova.virt.hardware [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1257.143010] env[65522]: DEBUG nova.virt.hardware [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1257.143229] env[65522]: DEBUG nova.virt.hardware [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1257.143387] env[65522]: DEBUG nova.virt.hardware [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1257.143546] env[65522]: DEBUG nova.virt.hardware [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1257.143703] env[65522]: DEBUG nova.virt.hardware [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1257.144086] env[65522]: DEBUG nova.virt.hardware [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1257.144990] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26de6335-c012-4769-9b83-a224b1ce2fd3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.153766] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c263a38-fec0-4ae4-92f5-78170b578f02 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.168096] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1257.173769] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1257.174065] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1257.174407] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-85cfe6f4-ee70-406d-9cf8-c78c76d0457c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.187574] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1257.193840] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1257.193840] env[65522]: value = "task-5115131" [ 1257.193840] env[65522]: _type = "Task" [ 1257.193840] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.202622] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115131, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.245159] env[65522]: DEBUG oslo_vmware.api [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115130, 'name': PowerOffVM_Task, 'duration_secs': 0.25076} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.245414] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1257.245587] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1257.247401] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c3e327ca-2513-466e-9300-4dd93d75621e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.328432] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1257.328803] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1257.329022] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Deleting the datastore file [datastore1] 3987d71f-47ab-4dd4-8426-95d78503effa {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1257.329370] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2865e83e-c47b-4236-b5c3-76a052b9505a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.338129] env[65522]: DEBUG oslo_vmware.api [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for the task: (returnval){ [ 1257.338129] env[65522]: value = "task-5115133" [ 1257.338129] env[65522]: _type = "Task" [ 1257.338129] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.348208] env[65522]: DEBUG oslo_vmware.api [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115133, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.383837] env[65522]: DEBUG oslo_concurrency.lockutils [None req-064ec591-c140-4a13-bfc2-9bec9f2e78a1 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.058s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1257.518676] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.237s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1257.518825] env[65522]: INFO nova.compute.manager [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Migrating [ 1257.526560] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.339s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1257.526819] env[65522]: DEBUG nova.objects.instance [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lazy-loading 'resources' on Instance uuid 874b6a07-ebba-4f75-937e-2281af1598b8 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1257.705142] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115131, 'name': CreateVM_Task, 'duration_secs': 0.369779} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.705291] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1257.705680] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1257.705858] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1257.706563] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1257.706563] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-627952e8-b75e-4e6d-a149-7d0b16a3530f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.712815] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1257.712815] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527454df-76f9-a730-8f8c-b1893ce806b1" [ 1257.712815] env[65522]: _type = "Task" [ 1257.712815] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.721732] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527454df-76f9-a730-8f8c-b1893ce806b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.756247] env[65522]: DEBUG nova.objects.instance [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lazy-loading 'flavor' on Instance uuid 1b0fcdbb-aa9e-4915-a8c6-1167aa888390 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1257.850133] env[65522]: DEBUG oslo_vmware.api [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Task: {'id': task-5115133, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216123} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.850372] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1257.850522] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1257.850698] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1257.850869] env[65522]: INFO nova.compute.manager [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1257.851183] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1257.851414] env[65522]: DEBUG nova.compute.manager [-] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1257.851534] env[65522]: DEBUG nova.network.neutron [-] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1257.851838] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1257.852443] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1257.852705] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1257.891304] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1258.039630] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1258.039839] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1258.039960] env[65522]: DEBUG nova.network.neutron [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1258.200808] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43b7962-8bc7-4220-a666-6dac7f3bb4e0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.210222] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e15e07-544a-4d50-b137-224ae788df83 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.262681] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]527454df-76f9-a730-8f8c-b1893ce806b1, 'name': SearchDatastore_Task, 'duration_secs': 0.011221} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.265266] env[65522]: DEBUG oslo_concurrency.lockutils [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1258.265502] env[65522]: DEBUG oslo_concurrency.lockutils [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1258.265683] env[65522]: DEBUG nova.network.neutron [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1258.265859] env[65522]: DEBUG nova.objects.instance [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lazy-loading 'info_cache' on Instance uuid 1b0fcdbb-aa9e-4915-a8c6-1167aa888390 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1258.267119] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1258.267344] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1258.267559] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1258.267698] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1258.267866] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1258.268904] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe5a8ed-1d8a-48eb-b0b6-901c1d0f1bcd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.273057] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-311b111b-1437-431d-a8c5-ed1e114198b1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.282470] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e71851-4f38-492f-87b6-3c51664303c9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.287951] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1258.288150] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1258.289406] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ad5aeec-8335-4702-8b05-9877e357f80a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.301267] env[65522]: DEBUG nova.compute.provider_tree [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1258.306772] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1258.306772] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b0128b-e3b9-799b-6f17-c578229e6e92" [ 1258.306772] env[65522]: _type = "Task" [ 1258.306772] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.315013] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b0128b-e3b9-799b-6f17-c578229e6e92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.400342] env[65522]: DEBUG nova.compute.manager [req-fd603c8d-6c6b-443a-92a4-4cbfbd8d6ca4 req-4f645e5f-23d1-46ca-bf5a-c38c17ac9e21 service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Received event network-vif-deleted-c8c5ef6a-c86a-4a39-a2cc-918754e4e78a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1258.400611] env[65522]: INFO nova.compute.manager [req-fd603c8d-6c6b-443a-92a4-4cbfbd8d6ca4 req-4f645e5f-23d1-46ca-bf5a-c38c17ac9e21 service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Neutron deleted interface c8c5ef6a-c86a-4a39-a2cc-918754e4e78a; detaching it from the instance and deleting it from the info cache [ 1258.400700] env[65522]: DEBUG nova.network.neutron [req-fd603c8d-6c6b-443a-92a4-4cbfbd8d6ca4 req-4f645e5f-23d1-46ca-bf5a-c38c17ac9e21 service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1258.543383] env[65522]: WARNING neutronclient.v2_0.client [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1258.544295] env[65522]: WARNING openstack [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1258.544759] env[65522]: WARNING openstack [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1258.685943] env[65522]: WARNING openstack [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1258.686352] env[65522]: WARNING openstack [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1258.748563] env[65522]: WARNING neutronclient.v2_0.client [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1258.749230] env[65522]: WARNING openstack [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1258.749575] env[65522]: WARNING openstack [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1258.774757] env[65522]: DEBUG nova.objects.base [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Object Instance<1b0fcdbb-aa9e-4915-a8c6-1167aa888390> lazy-loaded attributes: flavor,info_cache {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1258.804765] env[65522]: DEBUG nova.scheduler.client.report [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1258.817888] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b0128b-e3b9-799b-6f17-c578229e6e92, 'name': SearchDatastore_Task, 'duration_secs': 0.010238} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.818847] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb04873c-7696-4b6e-8fad-02986aafce91 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.825561] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1258.825561] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52340e14-546e-62ad-92e9-3f0a8d39a9d7" [ 1258.825561] env[65522]: _type = "Task" [ 1258.825561] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.832392] env[65522]: DEBUG nova.network.neutron [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance_info_cache with network_info: [{"id": "4a2d504c-073b-4b0a-b208-065c99308adc", "address": "fa:16:3e:81:96:f8", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a2d504c-07", "ovs_interfaceid": "4a2d504c-073b-4b0a-b208-065c99308adc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1258.836915] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52340e14-546e-62ad-92e9-3f0a8d39a9d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.871033] env[65522]: DEBUG nova.network.neutron [-] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1258.904071] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4d3cf50-646b-4965-81c1-5b9ea0a5a51f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.916382] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f11706-1196-4a83-8384-a74fe5d9161b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.949827] env[65522]: DEBUG nova.compute.manager [req-fd603c8d-6c6b-443a-92a4-4cbfbd8d6ca4 req-4f645e5f-23d1-46ca-bf5a-c38c17ac9e21 service nova] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Detach interface failed, port_id=c8c5ef6a-c86a-4a39-a2cc-918754e4e78a, reason: Instance 3987d71f-47ab-4dd4-8426-95d78503effa could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1259.278497] env[65522]: WARNING neutronclient.v2_0.client [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1259.279207] env[65522]: WARNING openstack [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1259.279587] env[65522]: WARNING openstack [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1259.309770] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.783s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1259.333475] env[65522]: INFO nova.scheduler.client.report [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Deleted allocations for instance 874b6a07-ebba-4f75-937e-2281af1598b8 [ 1259.339411] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1259.340173] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52340e14-546e-62ad-92e9-3f0a8d39a9d7, 'name': SearchDatastore_Task, 'duration_secs': 0.011205} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.342945] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1259.343046] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] c1aca457-dc7c-477c-9b4f-3ed577fb2adb/c1aca457-dc7c-477c-9b4f-3ed577fb2adb.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1259.343621] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0badc125-c187-447b-941f-5996d0a98c54 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.353281] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1259.353281] env[65522]: value = "task-5115134" [ 1259.353281] env[65522]: _type = "Task" [ 1259.353281] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.368287] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115134, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.372992] env[65522]: INFO nova.compute.manager [-] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Took 1.52 seconds to deallocate network for instance. [ 1259.407595] env[65522]: WARNING openstack [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1259.408017] env[65522]: WARNING openstack [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1259.494448] env[65522]: WARNING neutronclient.v2_0.client [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1259.495231] env[65522]: WARNING openstack [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1259.495567] env[65522]: WARNING openstack [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1259.590339] env[65522]: DEBUG nova.network.neutron [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Updating instance_info_cache with network_info: [{"id": "0a9cf0d6-c5cb-4805-918d-ed092924fb2a", "address": "fa:16:3e:23:17:05", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a9cf0d6-c5", "ovs_interfaceid": "0a9cf0d6-c5cb-4805-918d-ed092924fb2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1259.849012] env[65522]: DEBUG oslo_concurrency.lockutils [None req-c15f48de-402b-4ea5-b254-84a8f045ac96 tempest-ServerRescueTestJSON-76313780 tempest-ServerRescueTestJSON-76313780-project-member] Lock "874b6a07-ebba-4f75-937e-2281af1598b8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.064s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1259.864735] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115134, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504021} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.865968] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] c1aca457-dc7c-477c-9b4f-3ed577fb2adb/c1aca457-dc7c-477c-9b4f-3ed577fb2adb.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1259.865968] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1259.865968] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84327aec-d686-4fdb-8092-a58f090c9cfe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.873863] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1259.873863] env[65522]: value = "task-5115135" [ 1259.873863] env[65522]: _type = "Task" [ 1259.873863] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.880245] env[65522]: DEBUG oslo_concurrency.lockutils [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1259.880523] env[65522]: DEBUG oslo_concurrency.lockutils [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1259.880716] env[65522]: DEBUG nova.objects.instance [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lazy-loading 'resources' on Instance uuid 3987d71f-47ab-4dd4-8426-95d78503effa {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1259.886121] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115135, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.093717] env[65522]: DEBUG oslo_concurrency.lockutils [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1260.384114] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115135, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075317} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.384330] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1260.385355] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bbeb974-2bbb-4f18-8208-16aaa70a087c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.412320] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] c1aca457-dc7c-477c-9b4f-3ed577fb2adb/c1aca457-dc7c-477c-9b4f-3ed577fb2adb.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1260.415459] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8c3eab0-700e-4cce-b28c-e6be4a0a693d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.438059] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1260.438059] env[65522]: value = "task-5115136" [ 1260.438059] env[65522]: _type = "Task" [ 1260.438059] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.448355] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115136, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.549886] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686dafa1-7d4f-4230-8089-b10f289ac6d3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.558193] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c74f1c88-0f5a-4e3a-9b8d-4c1b7ad42691 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.591379] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c820a51-47ec-44ff-b0f1-dc0612eec534 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.601562] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1fe88a-1964-4f0e-8198-9c24ef851992 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.616159] env[65522]: DEBUG nova.compute.provider_tree [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1260.857375] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af7ede40-bfff-4c0c-8ae8-3a6eb1b9d372 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.880081] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance '3adfefa5-fb19-44b2-b3c5-42f2e2918673' progress to 0 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1260.948693] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115136, 'name': ReconfigVM_Task, 'duration_secs': 0.285051} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.948975] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Reconfigured VM instance instance-00000077 to attach disk [datastore2] c1aca457-dc7c-477c-9b4f-3ed577fb2adb/c1aca457-dc7c-477c-9b4f-3ed577fb2adb.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1260.949625] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4f640f87-c238-4a3b-9cd0-53884082ae27 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.956139] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1260.956139] env[65522]: value = "task-5115137" [ 1260.956139] env[65522]: _type = "Task" [ 1260.956139] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.967557] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115137, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.106891] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1261.106891] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-717fc9b6-d608-4574-bb1d-093c49dfdeb3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.114199] env[65522]: DEBUG oslo_vmware.api [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1261.114199] env[65522]: value = "task-5115138" [ 1261.114199] env[65522]: _type = "Task" [ 1261.114199] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.119672] env[65522]: DEBUG nova.scheduler.client.report [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1261.128806] env[65522]: DEBUG oslo_vmware.api [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115138, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.386708] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1261.386980] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4bf91405-5eb4-40c8-b551-49ea35d827c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.397870] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1261.397870] env[65522]: value = "task-5115139" [ 1261.397870] env[65522]: _type = "Task" [ 1261.397870] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.408451] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115139, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.466563] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115137, 'name': Rename_Task, 'duration_secs': 0.14749} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.466914] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1261.467135] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1e018bd-f4fe-48a3-987c-453bb7031e39 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.474959] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1261.474959] env[65522]: value = "task-5115140" [ 1261.474959] env[65522]: _type = "Task" [ 1261.474959] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.494117] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115140, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.624035] env[65522]: DEBUG oslo_vmware.api [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115138, 'name': PowerOnVM_Task, 'duration_secs': 0.390915} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.624399] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1261.624718] env[65522]: DEBUG nova.compute.manager [None req-932f2514-e4c2-4354-bd31-0c9b6f1473fb tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1261.625533] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-953db359-9251-443a-8624-ece0e52ee539 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.630008] env[65522]: DEBUG oslo_concurrency.lockutils [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.749s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1261.660290] env[65522]: INFO nova.scheduler.client.report [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Deleted allocations for instance 3987d71f-47ab-4dd4-8426-95d78503effa [ 1261.912879] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115139, 'name': PowerOffVM_Task, 'duration_secs': 0.208797} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.913348] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1261.913609] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance '3adfefa5-fb19-44b2-b3c5-42f2e2918673' progress to 17 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1261.987063] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115140, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.168219] env[65522]: DEBUG oslo_concurrency.lockutils [None req-25dab551-8ae0-42ac-8353-a4b9ef717792 tempest-AttachVolumeTestJSON-1670241509 tempest-AttachVolumeTestJSON-1670241509-project-member] Lock "3987d71f-47ab-4dd4-8426-95d78503effa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.959s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1262.421980] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1262.422292] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1262.422453] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1262.422635] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1262.422840] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1262.423060] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1262.423439] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1262.423670] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1262.423939] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1262.424253] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1262.424492] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1262.430598] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e00d10e-7bd1-4e39-8072-3d9ac7c3c6b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.449831] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1262.449831] env[65522]: value = "task-5115141" [ 1262.449831] env[65522]: _type = "Task" [ 1262.449831] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.460579] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115141, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.487603] env[65522]: DEBUG oslo_vmware.api [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115140, 'name': PowerOnVM_Task, 'duration_secs': 0.54969} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.487603] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1262.488043] env[65522]: DEBUG nova.compute.manager [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1262.488712] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b9e4ae-6fbd-4550-98ac-7dc8173ba230 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.661380] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60ee699-433e-402b-b0a4-7ddd97216b72 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.669770] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0777dff2-5b74-4d55-9632-db880eafc520 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Suspending the VM {{(pid=65522) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1262.670273] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-516c81e0-fc76-4ab4-938b-ba339a52016b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.677372] env[65522]: DEBUG oslo_vmware.api [None req-0777dff2-5b74-4d55-9632-db880eafc520 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1262.677372] env[65522]: value = "task-5115142" [ 1262.677372] env[65522]: _type = "Task" [ 1262.677372] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.685991] env[65522]: DEBUG oslo_vmware.api [None req-0777dff2-5b74-4d55-9632-db880eafc520 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115142, 'name': SuspendVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.961229] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115141, 'name': ReconfigVM_Task, 'duration_secs': 0.203881} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.961581] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance '3adfefa5-fb19-44b2-b3c5-42f2e2918673' progress to 33 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1263.008377] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1263.008726] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1263.009561] env[65522]: DEBUG nova.objects.instance [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1263.189400] env[65522]: DEBUG oslo_vmware.api [None req-0777dff2-5b74-4d55-9632-db880eafc520 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115142, 'name': SuspendVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.468723] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1263.468990] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1263.469194] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1263.469419] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1263.469566] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1263.469710] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1263.469915] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1263.470086] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1263.470262] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1263.470444] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1263.470603] env[65522]: DEBUG nova.virt.hardware [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1263.476395] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Reconfiguring VM instance instance-00000073 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1263.477150] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4bcf6450-70ed-486a-9c52-63fda9c79cf5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.498031] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1263.498031] env[65522]: value = "task-5115144" [ 1263.498031] env[65522]: _type = "Task" [ 1263.498031] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.507632] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115144, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.688457] env[65522]: DEBUG oslo_vmware.api [None req-0777dff2-5b74-4d55-9632-db880eafc520 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115142, 'name': SuspendVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.897394] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquiring lock "c1aca457-dc7c-477c-9b4f-3ed577fb2adb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1263.897662] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Lock "c1aca457-dc7c-477c-9b4f-3ed577fb2adb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1263.897875] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquiring lock "c1aca457-dc7c-477c-9b4f-3ed577fb2adb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1263.898067] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Lock "c1aca457-dc7c-477c-9b4f-3ed577fb2adb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1263.898270] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Lock "c1aca457-dc7c-477c-9b4f-3ed577fb2adb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1263.900648] env[65522]: INFO nova.compute.manager [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Terminating instance [ 1264.010867] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115144, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.018269] env[65522]: DEBUG oslo_concurrency.lockutils [None req-db15e88b-1fc3-4c8a-9d8d-d4033d1c8ea2 tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1264.189587] env[65522]: DEBUG oslo_vmware.api [None req-0777dff2-5b74-4d55-9632-db880eafc520 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115142, 'name': SuspendVM_Task, 'duration_secs': 1.338003} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.189806] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0777dff2-5b74-4d55-9632-db880eafc520 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Suspended the VM {{(pid=65522) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1264.189984] env[65522]: DEBUG nova.compute.manager [None req-0777dff2-5b74-4d55-9632-db880eafc520 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1264.190773] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6f0e31-ad86-4e27-a149-addc7830437f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.404399] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquiring lock "refresh_cache-c1aca457-dc7c-477c-9b4f-3ed577fb2adb" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1264.404594] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquired lock "refresh_cache-c1aca457-dc7c-477c-9b4f-3ed577fb2adb" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1264.404768] env[65522]: DEBUG nova.network.neutron [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1264.510479] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115144, 'name': ReconfigVM_Task, 'duration_secs': 0.526485} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.510884] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Reconfigured VM instance instance-00000073 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1264.511869] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25317afa-e0d5-4544-9dd4-98f4d135e47c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.540085] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] 3adfefa5-fb19-44b2-b3c5-42f2e2918673/3adfefa5-fb19-44b2-b3c5-42f2e2918673.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1264.541728] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ece427f-05b6-41d2-8d7e-d5a48b77cf15 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.566848] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1264.566848] env[65522]: value = "task-5115145" [ 1264.566848] env[65522]: _type = "Task" [ 1264.566848] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.577793] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115145, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.908578] env[65522]: WARNING neutronclient.v2_0.client [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1264.909054] env[65522]: WARNING openstack [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1264.909411] env[65522]: WARNING openstack [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1264.934368] env[65522]: DEBUG nova.network.neutron [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1264.995974] env[65522]: DEBUG nova.network.neutron [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1265.060841] env[65522]: INFO nova.compute.manager [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Resuming [ 1265.061482] env[65522]: DEBUG nova.objects.instance [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lazy-loading 'flavor' on Instance uuid 1b0fcdbb-aa9e-4915-a8c6-1167aa888390 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1265.077704] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115145, 'name': ReconfigVM_Task, 'duration_secs': 0.398524} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.077989] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Reconfigured VM instance instance-00000073 to attach disk [datastore1] 3adfefa5-fb19-44b2-b3c5-42f2e2918673/3adfefa5-fb19-44b2-b3c5-42f2e2918673.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1265.078323] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance '3adfefa5-fb19-44b2-b3c5-42f2e2918673' progress to 50 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1265.498940] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Releasing lock "refresh_cache-c1aca457-dc7c-477c-9b4f-3ed577fb2adb" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1265.499406] env[65522]: DEBUG nova.compute.manager [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1265.499609] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1265.500552] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8d4854-93dc-4d70-b5fa-983d044f0c9a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.509180] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1265.509467] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-732c108c-8ae0-4419-84d3-23ed4b18ec14 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.516412] env[65522]: DEBUG oslo_vmware.api [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1265.516412] env[65522]: value = "task-5115146" [ 1265.516412] env[65522]: _type = "Task" [ 1265.516412] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.525064] env[65522]: DEBUG oslo_vmware.api [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115146, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.586077] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113e6569-d79b-465a-9802-9494c2babafd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.616239] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9070417-9552-4a74-8178-9fa8512ef9c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.641574] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance '3adfefa5-fb19-44b2-b3c5-42f2e2918673' progress to 67 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1266.028127] env[65522]: DEBUG oslo_vmware.api [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115146, 'name': PowerOffVM_Task, 'duration_secs': 0.106562} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.028511] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1266.028704] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1266.029017] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1aaf4b2a-92ff-4ede-a651-cea2407eb39a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.054605] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1266.054959] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1266.055256] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Deleting the datastore file [datastore2] c1aca457-dc7c-477c-9b4f-3ed577fb2adb {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1266.055700] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e8b20bd8-f811-4b3a-beba-3eaa9a923ac1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.063648] env[65522]: DEBUG oslo_vmware.api [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for the task: (returnval){ [ 1266.063648] env[65522]: value = "task-5115149" [ 1266.063648] env[65522]: _type = "Task" [ 1266.063648] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.074093] env[65522]: DEBUG oslo_vmware.api [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115149, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.576747] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1266.577160] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquired lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1266.577160] env[65522]: DEBUG nova.network.neutron [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1266.578620] env[65522]: DEBUG oslo_vmware.api [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Task: {'id': task-5115149, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161639} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.579023] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1266.579279] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1266.579425] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1266.579547] env[65522]: INFO nova.compute.manager [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1266.579794] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1266.579993] env[65522]: DEBUG nova.compute.manager [-] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1266.580106] env[65522]: DEBUG nova.network.neutron [-] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1266.580349] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1266.580876] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1266.581144] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1266.604797] env[65522]: DEBUG nova.network.neutron [-] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1266.605106] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1267.081082] env[65522]: WARNING neutronclient.v2_0.client [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1267.081265] env[65522]: WARNING openstack [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1267.081544] env[65522]: WARNING openstack [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1267.107293] env[65522]: DEBUG nova.network.neutron [-] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1267.193121] env[65522]: WARNING openstack [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1267.193505] env[65522]: WARNING openstack [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1267.247841] env[65522]: WARNING neutronclient.v2_0.client [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1267.254025] env[65522]: WARNING neutronclient.v2_0.client [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1267.254723] env[65522]: WARNING openstack [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1267.255073] env[65522]: WARNING openstack [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1267.284080] env[65522]: DEBUG nova.network.neutron [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Port 4a2d504c-073b-4b0a-b208-065c99308adc binding to destination host cpu-1 is already ACTIVE {{(pid=65522) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 1267.340067] env[65522]: DEBUG nova.network.neutron [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Updating instance_info_cache with network_info: [{"id": "0a9cf0d6-c5cb-4805-918d-ed092924fb2a", "address": "fa:16:3e:23:17:05", "network": {"id": "f8aca55c-4152-4a66-8240-e5cb5efffe9c", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-980074746-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa11b46d9fe144f391233e6eb9c819d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a9cf0d6-c5", "ovs_interfaceid": "0a9cf0d6-c5cb-4805-918d-ed092924fb2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1267.610150] env[65522]: INFO nova.compute.manager [-] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Took 1.03 seconds to deallocate network for instance. [ 1267.843070] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Releasing lock "refresh_cache-1b0fcdbb-aa9e-4915-a8c6-1167aa888390" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1267.843950] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b9adfa-3837-469c-8b9d-bf9dc22badd3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.851513] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Resuming the VM {{(pid=65522) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1267.851763] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-393c8790-8681-4ccc-9046-192ffa045d1b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.859167] env[65522]: DEBUG oslo_vmware.api [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1267.859167] env[65522]: value = "task-5115150" [ 1267.859167] env[65522]: _type = "Task" [ 1267.859167] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.868258] env[65522]: DEBUG oslo_vmware.api [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115150, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.117040] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1268.117487] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1268.117729] env[65522]: DEBUG nova.objects.instance [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Lazy-loading 'resources' on Instance uuid c1aca457-dc7c-477c-9b4f-3ed577fb2adb {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1268.316964] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1268.317415] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1268.317749] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1268.374415] env[65522]: DEBUG oslo_vmware.api [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115150, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.741948] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fbd6c14-e68a-4d68-8235-f0cc23817f03 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.750729] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783b3835-9bc1-4064-912f-d03d39a7e253 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.783148] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dafe905-c7d5-4891-bc8b-9da18cd991af {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.791787] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e1a285-993b-4eae-b328-ba0cd412e650 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.806335] env[65522]: DEBUG nova.compute.provider_tree [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1268.871235] env[65522]: DEBUG oslo_vmware.api [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115150, 'name': PowerOnVM_Task, 'duration_secs': 0.581003} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.871566] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Resumed the VM {{(pid=65522) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1268.871808] env[65522]: DEBUG nova.compute.manager [None req-8721321d-f419-4bb8-9999-4daec785f759 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1268.872660] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf7d9ec-11ba-44dd-9546-e5d68be0bf51 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.309412] env[65522]: DEBUG nova.scheduler.client.report [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1269.325507] env[65522]: WARNING neutronclient.v2_0.client [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1269.357378] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1269.357655] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1269.357893] env[65522]: DEBUG nova.network.neutron [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1269.788082] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1269.788082] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1269.788671] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1269.788671] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1269.788671] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1269.791022] env[65522]: INFO nova.compute.manager [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Terminating instance [ 1269.815620] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.698s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1269.843564] env[65522]: INFO nova.scheduler.client.report [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Deleted allocations for instance c1aca457-dc7c-477c-9b4f-3ed577fb2adb [ 1269.860668] env[65522]: WARNING neutronclient.v2_0.client [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1269.861578] env[65522]: WARNING openstack [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1269.862074] env[65522]: WARNING openstack [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1270.003298] env[65522]: WARNING openstack [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1270.003844] env[65522]: WARNING openstack [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1270.068725] env[65522]: WARNING neutronclient.v2_0.client [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1270.069406] env[65522]: WARNING openstack [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1270.069757] env[65522]: WARNING openstack [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1270.148493] env[65522]: DEBUG nova.network.neutron [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance_info_cache with network_info: [{"id": "4a2d504c-073b-4b0a-b208-065c99308adc", "address": "fa:16:3e:81:96:f8", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a2d504c-07", "ovs_interfaceid": "4a2d504c-073b-4b0a-b208-065c99308adc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1270.295348] env[65522]: DEBUG nova.compute.manager [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1270.295581] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1270.296535] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90251daa-327a-4296-a437-b1dda5f36abc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.306017] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1270.306261] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a4fc56f-eac5-4260-8c48-31d6d098cb2e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.313065] env[65522]: DEBUG oslo_vmware.api [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1270.313065] env[65522]: value = "task-5115152" [ 1270.313065] env[65522]: _type = "Task" [ 1270.313065] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.321488] env[65522]: DEBUG oslo_vmware.api [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115152, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.352958] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5b3c6603-fe82-4ad3-820f-f1566cc33fdf tempest-ServersListShow298Test-1893024600 tempest-ServersListShow298Test-1893024600-project-member] Lock "c1aca457-dc7c-477c-9b4f-3ed577fb2adb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.455s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1270.652063] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1270.823766] env[65522]: DEBUG oslo_vmware.api [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115152, 'name': PowerOffVM_Task, 'duration_secs': 0.216646} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.824210] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1270.824210] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1270.824476] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c113643a-ecb4-4e36-82e8-60cf58db8cdd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.888675] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1270.888908] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1270.889110] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleting the datastore file [datastore1] 1b0fcdbb-aa9e-4915-a8c6-1167aa888390 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1270.889440] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf117d2a-5c31-4d19-89e4-3e673edbeb85 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.897184] env[65522]: DEBUG oslo_vmware.api [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for the task: (returnval){ [ 1270.897184] env[65522]: value = "task-5115154" [ 1270.897184] env[65522]: _type = "Task" [ 1270.897184] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.907544] env[65522]: DEBUG oslo_vmware.api [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115154, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.162196] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a56cf294-c44d-44fb-861b-b4fcbbf60f6e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.169629] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139f2c72-92d5-4a87-a02f-cdd2edeaad40 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.408765] env[65522]: DEBUG oslo_vmware.api [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Task: {'id': task-5115154, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168828} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.408765] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1271.408765] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1271.408933] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1271.409116] env[65522]: INFO nova.compute.manager [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1271.409365] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1271.409557] env[65522]: DEBUG nova.compute.manager [-] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1271.409650] env[65522]: DEBUG nova.network.neutron [-] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1271.409938] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1271.410971] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1271.410971] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1271.446495] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1271.917669] env[65522]: DEBUG nova.compute.manager [req-15ce06bd-ab2a-4870-915f-e4b6daf21009 req-4966a74a-21b6-45c2-a40d-3cf4b6b0097e service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Received event network-vif-deleted-0a9cf0d6-c5cb-4805-918d-ed092924fb2a {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1271.917931] env[65522]: INFO nova.compute.manager [req-15ce06bd-ab2a-4870-915f-e4b6daf21009 req-4966a74a-21b6-45c2-a40d-3cf4b6b0097e service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Neutron deleted interface 0a9cf0d6-c5cb-4805-918d-ed092924fb2a; detaching it from the instance and deleting it from the info cache [ 1271.918081] env[65522]: DEBUG nova.network.neutron [req-15ce06bd-ab2a-4870-915f-e4b6daf21009 req-4966a74a-21b6-45c2-a40d-3cf4b6b0097e service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1272.284245] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-875e0a64-5f37-4ad5-b16a-d23539888f67 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.309349] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc604b6-8423-4500-8db5-ac4680568c06 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.317008] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance '3adfefa5-fb19-44b2-b3c5-42f2e2918673' progress to 83 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1272.386542] env[65522]: DEBUG nova.network.neutron [-] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1272.421500] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f26249a-a3b9-4aff-b3dd-8a919c1aac03 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.432677] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a28952-53fb-4fbf-b978-3868a848f6f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.464460] env[65522]: DEBUG nova.compute.manager [req-15ce06bd-ab2a-4870-915f-e4b6daf21009 req-4966a74a-21b6-45c2-a40d-3cf4b6b0097e service nova] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Detach interface failed, port_id=0a9cf0d6-c5cb-4805-918d-ed092924fb2a, reason: Instance 1b0fcdbb-aa9e-4915-a8c6-1167aa888390 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1272.823361] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1272.823726] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f240a86b-e0da-4b06-a235-29adaf19a243 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.830960] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1272.830960] env[65522]: value = "task-5115155" [ 1272.830960] env[65522]: _type = "Task" [ 1272.830960] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.839860] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115155, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.890605] env[65522]: INFO nova.compute.manager [-] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Took 1.48 seconds to deallocate network for instance. [ 1273.340940] env[65522]: DEBUG oslo_vmware.api [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115155, 'name': PowerOnVM_Task, 'duration_secs': 0.410997} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1273.341308] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1273.341445] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34ac48a9-893a-4884-b96e-ca70bcf0c50a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance '3adfefa5-fb19-44b2-b3c5-42f2e2918673' progress to 100 {{(pid=65522) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1273.398496] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1273.398766] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1273.398981] env[65522]: DEBUG nova.objects.instance [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lazy-loading 'resources' on Instance uuid 1b0fcdbb-aa9e-4915-a8c6-1167aa888390 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1273.997713] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b60017-8a95-4f9f-a03c-7f910e16d605 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.006604] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b73e3e-1d41-4820-8ab5-cb44951960e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.040532] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1dd4b71-af77-4c75-aaa3-230325d7eee0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.049075] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc46cae-d6b8-4998-bf1d-959ef2a6c0ee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.064186] env[65522]: DEBUG nova.compute.provider_tree [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1274.568374] env[65522]: DEBUG nova.scheduler.client.report [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1275.073956] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.675s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1275.096286] env[65522]: INFO nova.scheduler.client.report [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Deleted allocations for instance 1b0fcdbb-aa9e-4915-a8c6-1167aa888390 [ 1275.432578] env[65522]: WARNING neutronclient.v2_0.client [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1275.467862] env[65522]: WARNING neutronclient.v2_0.client [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1275.468252] env[65522]: WARNING neutronclient.v2_0.client [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1275.501365] env[65522]: DEBUG nova.network.neutron [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Port 4a2d504c-073b-4b0a-b208-065c99308adc binding to destination host cpu-1 is already ACTIVE {{(pid=65522) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 1275.501651] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1275.501801] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1275.501964] env[65522]: DEBUG nova.network.neutron [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1275.604474] env[65522]: DEBUG oslo_concurrency.lockutils [None req-f2df71a4-ee29-473b-9717-da7042fba0d0 tempest-ServerActionsTestJSON-1355288579 tempest-ServerActionsTestJSON-1355288579-project-member] Lock "1b0fcdbb-aa9e-4915-a8c6-1167aa888390" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.816s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1276.004926] env[65522]: WARNING neutronclient.v2_0.client [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1276.005744] env[65522]: WARNING openstack [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1276.006129] env[65522]: WARNING openstack [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1276.201222] env[65522]: WARNING openstack [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1276.201222] env[65522]: WARNING openstack [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1276.357482] env[65522]: WARNING neutronclient.v2_0.client [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1276.358203] env[65522]: WARNING openstack [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1276.358553] env[65522]: WARNING openstack [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1276.439822] env[65522]: DEBUG nova.network.neutron [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance_info_cache with network_info: [{"id": "4a2d504c-073b-4b0a-b208-065c99308adc", "address": "fa:16:3e:81:96:f8", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a2d504c-07", "ovs_interfaceid": "4a2d504c-073b-4b0a-b208-065c99308adc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1276.943172] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1277.449865] env[65522]: DEBUG nova.compute.manager [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65522) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:924}} [ 1278.563978] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1278.563978] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1278.588616] env[65522]: DEBUG oslo_concurrency.lockutils [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "daffdd40-9ee3-4a80-b670-5772395a32ae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1278.588891] env[65522]: DEBUG oslo_concurrency.lockutils [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "daffdd40-9ee3-4a80-b670-5772395a32ae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1278.589165] env[65522]: DEBUG oslo_concurrency.lockutils [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "daffdd40-9ee3-4a80-b670-5772395a32ae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1278.589365] env[65522]: DEBUG oslo_concurrency.lockutils [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "daffdd40-9ee3-4a80-b670-5772395a32ae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1278.589611] env[65522]: DEBUG oslo_concurrency.lockutils [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "daffdd40-9ee3-4a80-b670-5772395a32ae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1278.591830] env[65522]: INFO nova.compute.manager [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Terminating instance [ 1278.614841] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1278.615150] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1278.615423] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1278.615628] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1278.615801] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1278.617829] env[65522]: INFO nova.compute.manager [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Terminating instance [ 1279.065596] env[65522]: DEBUG nova.objects.instance [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'migration_context' on Instance uuid 3adfefa5-fb19-44b2-b3c5-42f2e2918673 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1279.095858] env[65522]: DEBUG nova.compute.manager [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1279.096226] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1279.096746] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c76580a2-bdc0-44ae-9913-dd91be9437d4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.104370] env[65522]: DEBUG oslo_vmware.api [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1279.104370] env[65522]: value = "task-5115156" [ 1279.104370] env[65522]: _type = "Task" [ 1279.104370] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.113777] env[65522]: DEBUG oslo_vmware.api [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115156, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.121880] env[65522]: DEBUG nova.compute.manager [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1279.122209] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1279.123135] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082cc241-becb-4693-bf83-558c3ad0d291 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.130943] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1279.131266] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4941889f-fd34-4801-81e7-10679e383456 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.137767] env[65522]: DEBUG oslo_vmware.api [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1279.137767] env[65522]: value = "task-5115157" [ 1279.137767] env[65522]: _type = "Task" [ 1279.137767] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.147779] env[65522]: DEBUG oslo_vmware.api [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115157, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.590426] env[65522]: DEBUG nova.scheduler.client.report [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Refreshing inventories for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1279.610981] env[65522]: DEBUG nova.scheduler.client.report [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Updating ProviderTree inventory for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1279.611227] env[65522]: DEBUG nova.compute.provider_tree [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1279.619362] env[65522]: DEBUG oslo_vmware.api [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115156, 'name': PowerOffVM_Task, 'duration_secs': 0.203276} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.619673] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1279.619873] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1279.620075] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994961', 'volume_id': 'eacd0d18-4608-43ee-a54c-6655db952ccb', 'name': 'volume-eacd0d18-4608-43ee-a54c-6655db952ccb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'daffdd40-9ee3-4a80-b670-5772395a32ae', 'attached_at': '2025-12-12T19:09:05.000000', 'detached_at': '', 'volume_id': 'eacd0d18-4608-43ee-a54c-6655db952ccb', 'serial': 'eacd0d18-4608-43ee-a54c-6655db952ccb'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1279.620870] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ef3a0f-1de8-46b1-a50c-d4c5aa8cd81c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.624429] env[65522]: DEBUG nova.scheduler.client.report [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Refreshing aggregate associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, aggregates: None {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1279.646357] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1efd57de-a06d-4eb3-b36c-e3e54e267ea6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.649715] env[65522]: DEBUG nova.scheduler.client.report [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Refreshing trait associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1279.657274] env[65522]: DEBUG oslo_vmware.api [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115157, 'name': PowerOffVM_Task, 'duration_secs': 0.231385} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.659127] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1279.659266] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1279.659647] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-55a5b83e-6610-4845-b0ff-e8987e7b8249 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.661692] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72a642c9-e204-41ef-8865-7f9fb5379327 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.683473] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25beab88-785d-4a53-8cfb-007fecd25da1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.699566] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] The volume has not been displaced from its original location: [datastore1] volume-eacd0d18-4608-43ee-a54c-6655db952ccb/volume-eacd0d18-4608-43ee-a54c-6655db952ccb.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1279.705136] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Reconfiguring VM instance instance-00000074 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1279.708173] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f48a7679-6fee-43ad-8ef9-7d90ea6cbb19 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.728410] env[65522]: DEBUG oslo_vmware.api [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1279.728410] env[65522]: value = "task-5115159" [ 1279.728410] env[65522]: _type = "Task" [ 1279.728410] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.740437] env[65522]: DEBUG oslo_vmware.api [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115159, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.752753] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1279.752983] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1279.753189] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleting the datastore file [datastore2] 3b82f9ce-ea59-430d-adb6-918cc6cc48a4 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1279.753486] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d07ce0d-c803-43b7-a23e-35e373557171 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.760545] env[65522]: DEBUG oslo_vmware.api [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for the task: (returnval){ [ 1279.760545] env[65522]: value = "task-5115160" [ 1279.760545] env[65522]: _type = "Task" [ 1279.760545] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.772354] env[65522]: DEBUG oslo_vmware.api [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115160, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.787379] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc4d70f8-8071-41cf-bed6-fbd13de84091 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.794914] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56baa2ad-94da-4a8f-a956-9f3f36290c67 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.828041] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece511a5-d546-482c-a6d2-3729ebab620b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.836527] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52848c4f-9449-4551-8b6b-fc0156ffa65a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.851461] env[65522]: DEBUG nova.compute.provider_tree [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1280.240488] env[65522]: DEBUG oslo_vmware.api [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115159, 'name': ReconfigVM_Task, 'duration_secs': 0.178502} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.240787] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Reconfigured VM instance instance-00000074 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1280.246202] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76c92a79-050f-47b2-9777-2427b5c6f5b5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.262588] env[65522]: DEBUG oslo_vmware.api [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1280.262588] env[65522]: value = "task-5115161" [ 1280.262588] env[65522]: _type = "Task" [ 1280.262588] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.276670] env[65522]: DEBUG oslo_vmware.api [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Task: {'id': task-5115160, 'name': DeleteDatastoreFile_Task} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.280689] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1280.280916] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1280.281110] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1280.281290] env[65522]: INFO nova.compute.manager [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1280.281635] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1280.282069] env[65522]: DEBUG oslo_vmware.api [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115161, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.282069] env[65522]: DEBUG nova.compute.manager [-] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1280.282222] env[65522]: DEBUG nova.network.neutron [-] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1280.282434] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1280.282966] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1280.283231] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1280.351745] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1280.354458] env[65522]: DEBUG nova.scheduler.client.report [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1280.774535] env[65522]: DEBUG oslo_vmware.api [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115161, 'name': ReconfigVM_Task, 'duration_secs': 0.320388} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.774867] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994961', 'volume_id': 'eacd0d18-4608-43ee-a54c-6655db952ccb', 'name': 'volume-eacd0d18-4608-43ee-a54c-6655db952ccb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'daffdd40-9ee3-4a80-b670-5772395a32ae', 'attached_at': '2025-12-12T19:09:05.000000', 'detached_at': '', 'volume_id': 'eacd0d18-4608-43ee-a54c-6655db952ccb', 'serial': 'eacd0d18-4608-43ee-a54c-6655db952ccb'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1280.774867] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1280.776413] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc15f51-a32c-4a20-979a-e1ade331fcb4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.784818] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1280.785114] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e49d43b-f5c9-4d9a-935b-734858379478 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.850992] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1280.852789] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1280.852789] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleting the datastore file [datastore1] daffdd40-9ee3-4a80-b670-5772395a32ae {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1280.853500] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-175cbecb-19bb-4935-964c-fcd18e860bfc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.864892] env[65522]: DEBUG oslo_vmware.api [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1280.864892] env[65522]: value = "task-5115163" [ 1280.864892] env[65522]: _type = "Task" [ 1280.864892] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.870801] env[65522]: DEBUG nova.compute.manager [req-01d58996-08a8-4601-910b-1c2b645832f5 req-1a4eacf0-d62d-4423-b356-ce5eefd52e81 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Received event network-vif-deleted-fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1280.870986] env[65522]: INFO nova.compute.manager [req-01d58996-08a8-4601-910b-1c2b645832f5 req-1a4eacf0-d62d-4423-b356-ce5eefd52e81 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Neutron deleted interface fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc; detaching it from the instance and deleting it from the info cache [ 1280.871307] env[65522]: DEBUG nova.network.neutron [req-01d58996-08a8-4601-910b-1c2b645832f5 req-1a4eacf0-d62d-4423-b356-ce5eefd52e81 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1280.879822] env[65522]: DEBUG oslo_vmware.api [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115163, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.306033] env[65522]: DEBUG nova.network.neutron [-] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1281.365431] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.802s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1281.381177] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ea38998a-8416-4415-8f33-8c48281f5e5e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.390632] env[65522]: DEBUG oslo_vmware.api [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115163, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12211} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.392022] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1281.392022] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1281.392168] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1281.392362] env[65522]: INFO nova.compute.manager [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Took 2.30 seconds to destroy the instance on the hypervisor. [ 1281.392637] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1281.392926] env[65522]: DEBUG nova.compute.manager [-] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1281.393072] env[65522]: DEBUG nova.network.neutron [-] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1281.393268] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1281.393809] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1281.394079] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1281.403664] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139d76a0-8745-424e-a2ac-a9a81fa3800c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.435774] env[65522]: DEBUG nova.compute.manager [req-01d58996-08a8-4601-910b-1c2b645832f5 req-1a4eacf0-d62d-4423-b356-ce5eefd52e81 service nova] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Detach interface failed, port_id=fed73ed8-c0b2-4f6b-a3cd-0b813b3919fc, reason: Instance 3b82f9ce-ea59-430d-adb6-918cc6cc48a4 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1281.451481] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1281.784124] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquiring lock "bef9977d-ae21-426e-97d3-7b70a7276356" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1281.784487] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Lock "bef9977d-ae21-426e-97d3-7b70a7276356" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1281.808747] env[65522]: INFO nova.compute.manager [-] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Took 1.53 seconds to deallocate network for instance. [ 1282.287455] env[65522]: DEBUG nova.compute.manager [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1282.317248] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1282.317602] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1282.317872] env[65522]: DEBUG nova.objects.instance [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lazy-loading 'resources' on Instance uuid 3b82f9ce-ea59-430d-adb6-918cc6cc48a4 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1282.436928] env[65522]: DEBUG nova.network.neutron [-] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1282.810795] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1282.909506] env[65522]: INFO nova.compute.manager [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Swapping old allocation on dict_keys(['9a0d1ac8-ed26-4657-8899-fbceab29db0d']) held by migration 9498bd7f-755a-48dd-81bd-f3fcf24eea86 for instance [ 1282.917538] env[65522]: DEBUG nova.compute.manager [req-eab24ada-d999-48e3-ab51-04f0990a39a0 req-3d3c4830-d03f-4454-acf3-05362f8c0fc0 service nova] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Received event network-vif-deleted-6b80135d-6228-4366-83aa-ff70808ae005 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1282.933077] env[65522]: DEBUG nova.scheduler.client.report [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Overwriting current allocation {'allocations': {'9a0d1ac8-ed26-4657-8899-fbceab29db0d': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 165}}, 'project_id': '0d1239b79ae94cceb89ae7a8bd57da08', 'user_id': 'b51f1b47ad744dc2b2b0598cee11bc6f', 'consumer_generation': 1} on consumer 3adfefa5-fb19-44b2-b3c5-42f2e2918673 {{(pid=65522) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1282.936824] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0eb7bb8-5082-490c-8f23-f893886f616f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.939582] env[65522]: INFO nova.compute.manager [-] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Took 1.55 seconds to deallocate network for instance. [ 1282.947805] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4c453b-b535-4514-a1d1-e7c2d4dcbfcf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.980314] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3444073-51cb-4ffd-bfd7-f692dd6c2ed5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.985352] env[65522]: WARNING neutronclient.v2_0.client [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1282.991065] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a084266c-35f7-4559-a935-996c77fd09f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.004989] env[65522]: DEBUG nova.compute.provider_tree [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1283.022187] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.022376] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1283.022552] env[65522]: DEBUG nova.network.neutron [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1283.486359] env[65522]: INFO nova.compute.manager [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Took 0.55 seconds to detach 1 volumes for instance. [ 1283.488736] env[65522]: DEBUG nova.compute.manager [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Deleting volume: eacd0d18-4608-43ee-a54c-6655db952ccb {{(pid=65522) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3319}} [ 1283.508947] env[65522]: DEBUG nova.scheduler.client.report [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1283.525414] env[65522]: WARNING neutronclient.v2_0.client [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1283.526118] env[65522]: WARNING openstack [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.526475] env[65522]: WARNING openstack [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.667669] env[65522]: WARNING openstack [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.668083] env[65522]: WARNING openstack [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.729382] env[65522]: WARNING neutronclient.v2_0.client [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1283.730051] env[65522]: WARNING openstack [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.730405] env[65522]: WARNING openstack [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.811467] env[65522]: DEBUG nova.network.neutron [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance_info_cache with network_info: [{"id": "4a2d504c-073b-4b0a-b208-065c99308adc", "address": "fa:16:3e:81:96:f8", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a2d504c-07", "ovs_interfaceid": "4a2d504c-073b-4b0a-b208-065c99308adc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1284.014071] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.696s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1284.017209] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.206s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1284.018619] env[65522]: INFO nova.compute.claims [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1284.037634] env[65522]: INFO nova.scheduler.client.report [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Deleted allocations for instance 3b82f9ce-ea59-430d-adb6-918cc6cc48a4 [ 1284.040221] env[65522]: DEBUG oslo_concurrency.lockutils [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1284.315040] env[65522]: DEBUG oslo_concurrency.lockutils [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "refresh_cache-3adfefa5-fb19-44b2-b3c5-42f2e2918673" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1284.316186] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0beeec32-352c-42b2-8f0d-dfc275632902 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.324091] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1901a274-c927-4634-bc1c-530faaee5894 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.545550] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1a35d7ba-cfce-4eef-8fcd-262c88a92fa0 tempest-AttachVolumeShelveTestJSON-1393334336 tempest-AttachVolumeShelveTestJSON-1393334336-project-member] Lock "3b82f9ce-ea59-430d-adb6-918cc6cc48a4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.930s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1285.114761] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23895e67-02b0-4f1e-819e-994bd8357a4b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.123370] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d17388-d39d-4d27-929b-e4b66b24cacd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.156133] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b93ffb-462c-4931-b899-11ea58a41c29 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.164140] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29fbef9f-d84e-4860-bc51-bad5a25363cb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.178854] env[65522]: DEBUG nova.compute.provider_tree [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1285.413053] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1285.413440] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3514d8a8-8159-4644-8b07-de48b2e51db6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.422077] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1285.422077] env[65522]: value = "task-5115165" [ 1285.422077] env[65522]: _type = "Task" [ 1285.422077] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.430965] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115165, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.682711] env[65522]: DEBUG nova.scheduler.client.report [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1285.932183] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115165, 'name': PowerOffVM_Task, 'duration_secs': 0.219213} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.932439] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1285.933135] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1285.933333] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1285.933486] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1285.933667] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1285.933809] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1285.933951] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1285.934167] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1285.934363] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1285.934535] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1285.934700] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1285.934873] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1285.939763] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bed68f9e-f9e4-438a-90c2-939fdeb1131a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.956194] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1285.956194] env[65522]: value = "task-5115166" [ 1285.956194] env[65522]: _type = "Task" [ 1285.956194] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.964609] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115166, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.188267] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.171s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1286.188822] env[65522]: DEBUG nova.compute.manager [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1286.191583] env[65522]: DEBUG oslo_concurrency.lockutils [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.151s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1286.191804] env[65522]: DEBUG oslo_concurrency.lockutils [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1286.215207] env[65522]: INFO nova.scheduler.client.report [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleted allocations for instance daffdd40-9ee3-4a80-b670-5772395a32ae [ 1286.467548] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115166, 'name': ReconfigVM_Task, 'duration_secs': 0.153766} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.468415] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728ee418-fc95-4e0d-821f-1d1eead29a1f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.490105] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1286.490369] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1286.490524] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1286.490708] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1286.490852] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1286.490997] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1286.491220] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1286.491379] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1286.491543] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1286.491701] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1286.491872] env[65522]: DEBUG nova.virt.hardware [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1286.492768] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c8bbc66-7edc-4bfd-9156-7363358925c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.498774] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1286.498774] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5257d296-57f7-cf50-858c-29c5072ece79" [ 1286.498774] env[65522]: _type = "Task" [ 1286.498774] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.506930] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5257d296-57f7-cf50-858c-29c5072ece79, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.696484] env[65522]: DEBUG nova.compute.utils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1286.697997] env[65522]: DEBUG nova.compute.manager [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1286.698223] env[65522]: DEBUG nova.network.neutron [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1286.698583] env[65522]: WARNING neutronclient.v2_0.client [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1286.698899] env[65522]: WARNING neutronclient.v2_0.client [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1286.699491] env[65522]: WARNING openstack [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1286.699828] env[65522]: WARNING openstack [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1286.723044] env[65522]: DEBUG oslo_concurrency.lockutils [None req-38a1b5ae-814a-44e5-a1d7-6b663729cdc5 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "daffdd40-9ee3-4a80-b670-5772395a32ae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.134s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1286.760018] env[65522]: DEBUG nova.policy [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d4cd2e476494a3b8bca78a23cec1eae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd30a4e7fde4842489a863e9a2c4cf8b0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1287.010741] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5257d296-57f7-cf50-858c-29c5072ece79, 'name': SearchDatastore_Task, 'duration_secs': 0.008536} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.016582] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Reconfiguring VM instance instance-00000073 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1287.017082] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ea33fc7-1c0e-4118-acfe-e873145e9e03 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.039263] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1287.039263] env[65522]: value = "task-5115168" [ 1287.039263] env[65522]: _type = "Task" [ 1287.039263] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.049301] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115168, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.077820] env[65522]: DEBUG nova.network.neutron [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Successfully created port: 3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1287.208910] env[65522]: DEBUG nova.compute.manager [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1287.552530] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115168, 'name': ReconfigVM_Task, 'duration_secs': 0.24507} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.552909] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Reconfigured VM instance instance-00000073 to detach disk 2000 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1287.553956] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9642ea6c-4bba-4999-9df9-8932df31b2a3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.583107] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] 3adfefa5-fb19-44b2-b3c5-42f2e2918673/3adfefa5-fb19-44b2-b3c5-42f2e2918673.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1287.583470] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7df18011-8fd8-41ed-ab37-e00395beb219 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.603599] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1287.603599] env[65522]: value = "task-5115169" [ 1287.603599] env[65522]: _type = "Task" [ 1287.603599] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.613358] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115169, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.792422] env[65522]: DEBUG oslo_concurrency.lockutils [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1287.792860] env[65522]: DEBUG oslo_concurrency.lockutils [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1287.793145] env[65522]: DEBUG oslo_concurrency.lockutils [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1287.793391] env[65522]: DEBUG oslo_concurrency.lockutils [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1287.793594] env[65522]: DEBUG oslo_concurrency.lockutils [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1287.796466] env[65522]: INFO nova.compute.manager [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Terminating instance [ 1288.114839] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115169, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.219940] env[65522]: DEBUG nova.compute.manager [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1288.249504] env[65522]: DEBUG nova.virt.hardware [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1288.249763] env[65522]: DEBUG nova.virt.hardware [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1288.249916] env[65522]: DEBUG nova.virt.hardware [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1288.250102] env[65522]: DEBUG nova.virt.hardware [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1288.250254] env[65522]: DEBUG nova.virt.hardware [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1288.250401] env[65522]: DEBUG nova.virt.hardware [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1288.250614] env[65522]: DEBUG nova.virt.hardware [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1288.250768] env[65522]: DEBUG nova.virt.hardware [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1288.251139] env[65522]: DEBUG nova.virt.hardware [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1288.251336] env[65522]: DEBUG nova.virt.hardware [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1288.251512] env[65522]: DEBUG nova.virt.hardware [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1288.252422] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c8fc74-74b8-412a-952f-08fd3ff64616 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.261272] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8be1f7-a63a-442a-8629-ada8cd816cac {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.300630] env[65522]: DEBUG nova.compute.manager [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1288.300892] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1288.301851] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf38fc5c-3a0f-4a0f-a9aa-fdc0737b5fb0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.311147] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1288.311463] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00b8649e-632e-4190-a701-c9c297b1ad48 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.319288] env[65522]: DEBUG oslo_vmware.api [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1288.319288] env[65522]: value = "task-5115170" [ 1288.319288] env[65522]: _type = "Task" [ 1288.319288] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.329724] env[65522]: DEBUG oslo_vmware.api [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115170, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.567612] env[65522]: DEBUG nova.compute.manager [req-9eea722f-f1cb-4e40-a63b-e9e9913af2e8 req-03aae13c-ffa9-4c85-8910-44f9d9b725ce service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Received event network-vif-plugged-3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1288.567852] env[65522]: DEBUG oslo_concurrency.lockutils [req-9eea722f-f1cb-4e40-a63b-e9e9913af2e8 req-03aae13c-ffa9-4c85-8910-44f9d9b725ce service nova] Acquiring lock "bef9977d-ae21-426e-97d3-7b70a7276356-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1288.568094] env[65522]: DEBUG oslo_concurrency.lockutils [req-9eea722f-f1cb-4e40-a63b-e9e9913af2e8 req-03aae13c-ffa9-4c85-8910-44f9d9b725ce service nova] Lock "bef9977d-ae21-426e-97d3-7b70a7276356-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1288.568255] env[65522]: DEBUG oslo_concurrency.lockutils [req-9eea722f-f1cb-4e40-a63b-e9e9913af2e8 req-03aae13c-ffa9-4c85-8910-44f9d9b725ce service nova] Lock "bef9977d-ae21-426e-97d3-7b70a7276356-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1288.568423] env[65522]: DEBUG nova.compute.manager [req-9eea722f-f1cb-4e40-a63b-e9e9913af2e8 req-03aae13c-ffa9-4c85-8910-44f9d9b725ce service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] No waiting events found dispatching network-vif-plugged-3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1288.568590] env[65522]: WARNING nova.compute.manager [req-9eea722f-f1cb-4e40-a63b-e9e9913af2e8 req-03aae13c-ffa9-4c85-8910-44f9d9b725ce service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Received unexpected event network-vif-plugged-3a8642fb-edda-4cd0-8293-86da0093bee5 for instance with vm_state building and task_state spawning. [ 1288.615056] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115169, 'name': ReconfigVM_Task, 'duration_secs': 0.932676} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.615169] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Reconfigured VM instance instance-00000073 to attach disk [datastore1] 3adfefa5-fb19-44b2-b3c5-42f2e2918673/3adfefa5-fb19-44b2-b3c5-42f2e2918673.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1288.616190] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c376df3-61b3-4c78-95d6-cbcf481c8334 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.644746] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abed6ade-4eb9-4f78-9455-888d51da8a8d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.666905] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0893df3d-df33-4066-81ee-87fd78c122fa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.689016] env[65522]: DEBUG nova.network.neutron [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Successfully updated port: 3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1288.690922] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ab7af4-a626-4ac3-a0f9-e618d16f74cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.700461] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1288.701457] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53130951-19b5-49ae-b8e3-54c338b43b9e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.709637] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1288.709637] env[65522]: value = "task-5115171" [ 1288.709637] env[65522]: _type = "Task" [ 1288.709637] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.719937] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115171, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.739222] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1288.739598] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1288.831811] env[65522]: DEBUG oslo_vmware.api [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115170, 'name': PowerOffVM_Task, 'duration_secs': 0.206467} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.832098] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1288.832271] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1288.832542] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c0d93c7a-5329-4fba-9511-c5728ed72561 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.899066] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1288.899487] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1288.899720] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleting the datastore file [datastore1] aebb7082-f88c-42f3-9d9f-93cb2a6f545b {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1288.900058] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4f7424d-d08b-430b-aafb-9c11212eeb33 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.907418] env[65522]: DEBUG oslo_vmware.api [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1288.907418] env[65522]: value = "task-5115173" [ 1288.907418] env[65522]: _type = "Task" [ 1288.907418] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.916315] env[65522]: DEBUG oslo_vmware.api [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115173, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.195952] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquiring lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1289.196215] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquired lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1289.196372] env[65522]: DEBUG nova.network.neutron [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1289.219724] env[65522]: DEBUG oslo_vmware.api [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115171, 'name': PowerOnVM_Task, 'duration_secs': 0.4238} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.219997] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1289.245070] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1289.245842] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1289.245842] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1289.245842] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1289.245842] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1289.245842] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1289.246029] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1289.246360] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1289.420070] env[65522]: DEBUG oslo_vmware.api [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115173, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141908} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.420374] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1289.420562] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1289.420735] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1289.420902] env[65522]: INFO nova.compute.manager [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1289.421175] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1289.421377] env[65522]: DEBUG nova.compute.manager [-] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1289.421477] env[65522]: DEBUG nova.network.neutron [-] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1289.421846] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1289.422435] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1289.422693] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1289.462687] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1289.699835] env[65522]: WARNING openstack [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1289.700264] env[65522]: WARNING openstack [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1289.741656] env[65522]: DEBUG nova.network.neutron [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1289.748710] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1289.749356] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1289.749356] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1289.749356] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1289.750417] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5092c3f7-5adf-401d-80ff-0388d51cfc44 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.759416] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28837b61-9df4-4bbe-ae09-eabcb6cf4cb0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.765292] env[65522]: WARNING openstack [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1289.765685] env[65522]: WARNING openstack [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1289.784587] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6848f4e3-9cda-4f16-9a1b-7928ea0605f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.793403] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68a9fc2-079c-4bf3-9fee-2b8fd3be2ae3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.836382] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179290MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1289.836558] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1289.836826] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1289.917803] env[65522]: WARNING neutronclient.v2_0.client [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1289.918793] env[65522]: WARNING openstack [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1289.919375] env[65522]: WARNING openstack [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1290.036242] env[65522]: DEBUG nova.network.neutron [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Updating instance_info_cache with network_info: [{"id": "3a8642fb-edda-4cd0-8293-86da0093bee5", "address": "fa:16:3e:37:b0:01", "network": {"id": "6a449873-f579-41da-b6ae-a83934c5685a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1387860566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d30a4e7fde4842489a863e9a2c4cf8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "81d39ad2-4e62-4f09-a567-88ac5aa70467", "external-id": "nsx-vlan-transportzone-221", "segmentation_id": 221, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8642fb-ed", "ovs_interfaceid": "3a8642fb-edda-4cd0-8293-86da0093bee5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1290.187908] env[65522]: DEBUG nova.network.neutron [-] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1290.264486] env[65522]: INFO nova.compute.manager [None req-40ab62af-43fa-4e2a-af55-4f946fc1b23d tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance to original state: 'active' [ 1290.538740] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Releasing lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1290.539179] env[65522]: DEBUG nova.compute.manager [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Instance network_info: |[{"id": "3a8642fb-edda-4cd0-8293-86da0093bee5", "address": "fa:16:3e:37:b0:01", "network": {"id": "6a449873-f579-41da-b6ae-a83934c5685a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1387860566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d30a4e7fde4842489a863e9a2c4cf8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "81d39ad2-4e62-4f09-a567-88ac5aa70467", "external-id": "nsx-vlan-transportzone-221", "segmentation_id": 221, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8642fb-ed", "ovs_interfaceid": "3a8642fb-edda-4cd0-8293-86da0093bee5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1290.539675] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:37:b0:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '81d39ad2-4e62-4f09-a567-88ac5aa70467', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a8642fb-edda-4cd0-8293-86da0093bee5', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1290.548597] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Creating folder: Project (d30a4e7fde4842489a863e9a2c4cf8b0). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1290.549896] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-22644d69-2241-4852-85e7-76bafc8a37a7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.561990] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Created folder: Project (d30a4e7fde4842489a863e9a2c4cf8b0) in parent group-v994660. [ 1290.562315] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Creating folder: Instances. Parent ref: group-v994983. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1290.562607] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0737ad8f-1f25-4045-ba61-efe842ad0c48 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.573615] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Created folder: Instances in parent group-v994983. [ 1290.574097] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1290.574097] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1290.574407] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82faae95-8778-466b-9354-35792096bb30 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.599893] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1290.599893] env[65522]: value = "task-5115176" [ 1290.599893] env[65522]: _type = "Task" [ 1290.599893] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.608364] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115176, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.617888] env[65522]: DEBUG nova.compute.manager [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Received event network-changed-3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1290.617888] env[65522]: DEBUG nova.compute.manager [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Refreshing instance network info cache due to event network-changed-3a8642fb-edda-4cd0-8293-86da0093bee5. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1290.617888] env[65522]: DEBUG oslo_concurrency.lockutils [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] Acquiring lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1290.618175] env[65522]: DEBUG oslo_concurrency.lockutils [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] Acquired lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1290.618286] env[65522]: DEBUG nova.network.neutron [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Refreshing network info cache for port 3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1290.692034] env[65522]: INFO nova.compute.manager [-] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Took 1.27 seconds to deallocate network for instance. [ 1290.866464] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 7617189c-a902-42e7-8165-0e7c4a1de06d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.866628] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance aebb7082-f88c-42f3-9d9f-93cb2a6f545b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.866755] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance bef9977d-ae21-426e-97d3-7b70a7276356 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.866938] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 3adfefa5-fb19-44b2-b3c5-42f2e2918673 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1290.867544] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1290.867544] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=100GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '4', 'num_vm_active': '2', 'num_task_None': '1', 'num_os_type_None': '4', 'num_proj_91597233ae9c44c094f4c32d90332fa6': '2', 'io_workload': '1', 'num_task_deleting': '1', 'num_vm_resized': '1', 'num_task_resize_reverting': '1', 'num_proj_0d1239b79ae94cceb89ae7a8bd57da08': '1', 'num_vm_building': '1', 'num_task_spawning': '1', 'num_proj_d30a4e7fde4842489a863e9a2c4cf8b0': '1'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1290.936515] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc513c4b-0d07-44c5-903f-fdb4574d6c82 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.945330] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d6305a-d32f-4e57-bfa3-e8b77352e20f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.979646] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc19a173-5c70-4b68-9b26-32740955d18f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.988495] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e8aa3d-a381-4069-888c-07d1818d91c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.004208] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1291.109708] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115176, 'name': CreateVM_Task, 'duration_secs': 0.316752} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.109913] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1291.112527] env[65522]: WARNING neutronclient.v2_0.client [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1291.112527] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1291.112527] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1291.112527] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1291.112527] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be50f5d7-bc72-4d05-874a-52cc4c076996 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.117454] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1291.117454] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c44dbc-76d7-24e5-ed80-558552f06b98" [ 1291.117454] env[65522]: _type = "Task" [ 1291.117454] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.120985] env[65522]: WARNING neutronclient.v2_0.client [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1291.122560] env[65522]: WARNING openstack [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1291.122560] env[65522]: WARNING openstack [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1291.137494] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c44dbc-76d7-24e5-ed80-558552f06b98, 'name': SearchDatastore_Task, 'duration_secs': 0.010874} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.137764] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1291.137992] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1291.138230] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1291.138372] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1291.138551] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1291.138837] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-52d6f0b9-84bb-48fa-aad2-3bf543c0215e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.147948] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1291.148147] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1291.149510] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78bcd88d-6433-4b40-83dd-9ee264f3c95b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.155179] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1291.155179] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526c2ca7-80f3-a7a8-e678-fa9c186eafc0" [ 1291.155179] env[65522]: _type = "Task" [ 1291.155179] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.163189] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526c2ca7-80f3-a7a8-e678-fa9c186eafc0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.203429] env[65522]: DEBUG oslo_concurrency.lockutils [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1291.252319] env[65522]: WARNING openstack [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1291.252870] env[65522]: WARNING openstack [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1291.330629] env[65522]: WARNING neutronclient.v2_0.client [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1291.331337] env[65522]: WARNING openstack [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1291.331693] env[65522]: WARNING openstack [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1291.433965] env[65522]: DEBUG nova.network.neutron [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Updated VIF entry in instance network info cache for port 3a8642fb-edda-4cd0-8293-86da0093bee5. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1291.434383] env[65522]: DEBUG nova.network.neutron [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Updating instance_info_cache with network_info: [{"id": "3a8642fb-edda-4cd0-8293-86da0093bee5", "address": "fa:16:3e:37:b0:01", "network": {"id": "6a449873-f579-41da-b6ae-a83934c5685a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1387860566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d30a4e7fde4842489a863e9a2c4cf8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "81d39ad2-4e62-4f09-a567-88ac5aa70467", "external-id": "nsx-vlan-transportzone-221", "segmentation_id": 221, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8642fb-ed", "ovs_interfaceid": "3a8642fb-edda-4cd0-8293-86da0093bee5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1291.510327] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1291.669129] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526c2ca7-80f3-a7a8-e678-fa9c186eafc0, 'name': SearchDatastore_Task, 'duration_secs': 0.010189} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.670066] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-428e5ab3-eb5d-434f-91e4-bcfd16373fc2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.677143] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1291.677143] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f8d378-bcb7-7434-8041-70a3cb35aad2" [ 1291.677143] env[65522]: _type = "Task" [ 1291.677143] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.687166] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f8d378-bcb7-7434-8041-70a3cb35aad2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.928757] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1291.929047] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1291.929268] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1291.929450] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1291.929643] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1291.931969] env[65522]: INFO nova.compute.manager [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Terminating instance [ 1291.938790] env[65522]: DEBUG oslo_concurrency.lockutils [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] Releasing lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1291.939016] env[65522]: DEBUG nova.compute.manager [req-74092537-938a-4e66-a327-445f7a7b86e2 req-e6f0950c-dfe3-4835-97e9-74657fe7d19f service nova] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Received event network-vif-deleted-0b80621b-795f-4503-9161-c9799b6339b6 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1292.015559] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1292.015806] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.179s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1292.016163] env[65522]: DEBUG oslo_concurrency.lockutils [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.813s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1292.016450] env[65522]: DEBUG nova.objects.instance [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lazy-loading 'resources' on Instance uuid aebb7082-f88c-42f3-9d9f-93cb2a6f545b {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1292.188724] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f8d378-bcb7-7434-8041-70a3cb35aad2, 'name': SearchDatastore_Task, 'duration_secs': 0.026053} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.189032] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1292.189280] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] bef9977d-ae21-426e-97d3-7b70a7276356/bef9977d-ae21-426e-97d3-7b70a7276356.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1292.189564] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8d601f1c-2c7d-4b5c-a9f6-481081a7285f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.198145] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1292.198145] env[65522]: value = "task-5115177" [ 1292.198145] env[65522]: _type = "Task" [ 1292.198145] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.207465] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115177, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.436593] env[65522]: DEBUG nova.compute.manager [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1292.437057] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1292.437188] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e60de639-45dc-4b60-bf41-bf6fffff76f3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.447122] env[65522]: DEBUG oslo_vmware.api [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1292.447122] env[65522]: value = "task-5115178" [ 1292.447122] env[65522]: _type = "Task" [ 1292.447122] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.456966] env[65522]: DEBUG oslo_vmware.api [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115178, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.600107] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d2c73a-5eb9-4162-82d2-bfcf5b1d59b2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.612450] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-786220c8-e6c5-433d-a520-8faa61d85337 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.666506] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e8c961-29c3-4cc1-984d-121203b4b5af {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.680499] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83e4b09-82e5-4009-af73-461452bf03ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.707597] env[65522]: DEBUG nova.compute.provider_tree [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1292.724284] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115177, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.958468] env[65522]: DEBUG oslo_vmware.api [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115178, 'name': PowerOffVM_Task, 'duration_secs': 0.265486} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.958661] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1292.958859] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1292.959225] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994978', 'volume_id': '25a65c4f-a464-4ece-b85f-95bc2dfd3903', 'name': 'volume-25a65c4f-a464-4ece-b85f-95bc2dfd3903', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '3adfefa5-fb19-44b2-b3c5-42f2e2918673', 'attached_at': '2025-12-12T19:09:49.000000', 'detached_at': '', 'volume_id': '25a65c4f-a464-4ece-b85f-95bc2dfd3903', 'serial': '25a65c4f-a464-4ece-b85f-95bc2dfd3903'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1292.960087] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe34ef7-f13b-4012-b9ae-a2ef44133d01 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.982848] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9373e9e-af18-4ccd-b252-117bcc9070e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.992635] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb515bb2-c61c-47b3-94ad-113acc3d6e50 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.015167] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9206360a-5d63-4515-9959-c23be37690b8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.031985] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The volume has not been displaced from its original location: [datastore1] volume-25a65c4f-a464-4ece-b85f-95bc2dfd3903/volume-25a65c4f-a464-4ece-b85f-95bc2dfd3903.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1293.037491] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Reconfiguring VM instance instance-00000073 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1293.037851] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46555b37-6067-41ed-8c96-76310a3e3434 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.057590] env[65522]: DEBUG oslo_vmware.api [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1293.057590] env[65522]: value = "task-5115179" [ 1293.057590] env[65522]: _type = "Task" [ 1293.057590] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.067353] env[65522]: DEBUG oslo_vmware.api [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115179, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.217256] env[65522]: DEBUG nova.scheduler.client.report [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1293.220314] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115177, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.7242} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.220761] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] bef9977d-ae21-426e-97d3-7b70a7276356/bef9977d-ae21-426e-97d3-7b70a7276356.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1293.220981] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1293.221254] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7a648a3f-e48e-4949-bf88-1b8c9d105b45 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.228830] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1293.228830] env[65522]: value = "task-5115180" [ 1293.228830] env[65522]: _type = "Task" [ 1293.228830] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.239322] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115180, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.571391] env[65522]: DEBUG oslo_vmware.api [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115179, 'name': ReconfigVM_Task, 'duration_secs': 0.267334} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.571391] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Reconfigured VM instance instance-00000073 to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1293.576334] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3453ac1e-ecd4-421d-9b76-43606994cdde {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.594857] env[65522]: DEBUG oslo_vmware.api [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1293.594857] env[65522]: value = "task-5115181" [ 1293.594857] env[65522]: _type = "Task" [ 1293.594857] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.605121] env[65522]: DEBUG oslo_vmware.api [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115181, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.722299] env[65522]: DEBUG oslo_concurrency.lockutils [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.706s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1293.740943] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115180, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076386} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1293.741814] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1293.742691] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccffa354-0847-4e07-80bc-e62c6efa60e1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.748158] env[65522]: INFO nova.scheduler.client.report [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleted allocations for instance aebb7082-f88c-42f3-9d9f-93cb2a6f545b [ 1293.784105] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] bef9977d-ae21-426e-97d3-7b70a7276356/bef9977d-ae21-426e-97d3-7b70a7276356.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1293.784782] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-415090ea-a7df-4c3a-b5db-5f330e7299c8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.806848] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1293.806848] env[65522]: value = "task-5115182" [ 1293.806848] env[65522]: _type = "Task" [ 1293.806848] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.819589] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115182, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.105049] env[65522]: DEBUG oslo_vmware.api [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115181, 'name': ReconfigVM_Task, 'duration_secs': 0.341037} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.105306] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-994978', 'volume_id': '25a65c4f-a464-4ece-b85f-95bc2dfd3903', 'name': 'volume-25a65c4f-a464-4ece-b85f-95bc2dfd3903', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '3adfefa5-fb19-44b2-b3c5-42f2e2918673', 'attached_at': '2025-12-12T19:09:49.000000', 'detached_at': '', 'volume_id': '25a65c4f-a464-4ece-b85f-95bc2dfd3903', 'serial': '25a65c4f-a464-4ece-b85f-95bc2dfd3903'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1294.105622] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1294.106433] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31535e3e-9fe2-4606-aa68-b9fa15ebff18 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.113435] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1294.113681] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8498fa9b-e27f-4f67-88c6-ef1db8035fbe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.175793] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1294.176016] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1294.176225] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleting the datastore file [datastore1] 3adfefa5-fb19-44b2-b3c5-42f2e2918673 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1294.176518] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5aa8957a-2210-4eaa-a514-11c7b4de1f4d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.183964] env[65522]: DEBUG oslo_vmware.api [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1294.183964] env[65522]: value = "task-5115185" [ 1294.183964] env[65522]: _type = "Task" [ 1294.183964] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.192578] env[65522]: DEBUG oslo_vmware.api [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115185, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.288359] env[65522]: DEBUG oslo_concurrency.lockutils [None req-181adec2-4e71-43bc-bf17-d5b935dba9f4 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "aebb7082-f88c-42f3-9d9f-93cb2a6f545b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.495s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1294.317605] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115182, 'name': ReconfigVM_Task, 'duration_secs': 0.360252} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.317950] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Reconfigured VM instance instance-00000078 to attach disk [datastore2] bef9977d-ae21-426e-97d3-7b70a7276356/bef9977d-ae21-426e-97d3-7b70a7276356.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1294.318670] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3abc53a2-da66-4bce-a6c9-91b4a1c90ff5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.326751] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1294.326751] env[65522]: value = "task-5115186" [ 1294.326751] env[65522]: _type = "Task" [ 1294.326751] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.335568] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115186, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.695143] env[65522]: DEBUG oslo_vmware.api [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115185, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184039} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.695143] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1294.695564] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1294.695564] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1294.695649] env[65522]: INFO nova.compute.manager [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Took 2.26 seconds to destroy the instance on the hypervisor. [ 1294.695893] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1294.696120] env[65522]: DEBUG nova.compute.manager [-] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1294.696221] env[65522]: DEBUG nova.network.neutron [-] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1294.696464] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1294.696998] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1294.697274] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1294.742825] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1294.840988] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115186, 'name': Rename_Task, 'duration_secs': 0.152173} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1294.840988] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1294.840988] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6fcfc772-e105-4706-ba21-54c5c10d1e38 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1294.845895] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1294.845895] env[65522]: value = "task-5115187" [ 1294.845895] env[65522]: _type = "Task" [ 1294.845895] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1294.854222] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115187, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.189608] env[65522]: DEBUG nova.compute.manager [req-14c1b193-4d1a-4c91-8c10-d0b31bfcd760 req-86d4a9ea-a8a5-4c9a-b961-f0a94d536ce8 service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Received event network-vif-deleted-4a2d504c-073b-4b0a-b208-065c99308adc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1295.189809] env[65522]: INFO nova.compute.manager [req-14c1b193-4d1a-4c91-8c10-d0b31bfcd760 req-86d4a9ea-a8a5-4c9a-b961-f0a94d536ce8 service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Neutron deleted interface 4a2d504c-073b-4b0a-b208-065c99308adc; detaching it from the instance and deleting it from the info cache [ 1295.189979] env[65522]: DEBUG nova.network.neutron [req-14c1b193-4d1a-4c91-8c10-d0b31bfcd760 req-86d4a9ea-a8a5-4c9a-b961-f0a94d536ce8 service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1295.356447] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115187, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.666417] env[65522]: DEBUG nova.network.neutron [-] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1295.693760] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a03ca41-5c76-47dc-9b9c-422728023cba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.703675] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894b6feb-a510-4acf-bbc6-edae302603bf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.732865] env[65522]: DEBUG nova.compute.manager [req-14c1b193-4d1a-4c91-8c10-d0b31bfcd760 req-86d4a9ea-a8a5-4c9a-b961-f0a94d536ce8 service nova] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Detach interface failed, port_id=4a2d504c-073b-4b0a-b208-065c99308adc, reason: Instance 3adfefa5-fb19-44b2-b3c5-42f2e2918673 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1295.859025] env[65522]: DEBUG oslo_vmware.api [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115187, 'name': PowerOnVM_Task, 'duration_secs': 0.742172} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1295.859025] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1295.859025] env[65522]: INFO nova.compute.manager [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Took 7.64 seconds to spawn the instance on the hypervisor. [ 1295.859025] env[65522]: DEBUG nova.compute.manager [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1295.859025] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d69866-1221-4117-8a0e-f092bdc0c78c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1296.169767] env[65522]: INFO nova.compute.manager [-] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Took 1.47 seconds to deallocate network for instance. [ 1296.376242] env[65522]: INFO nova.compute.manager [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Took 13.58 seconds to build instance. [ 1296.711949] env[65522]: INFO nova.compute.manager [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Took 0.54 seconds to detach 1 volumes for instance. [ 1296.879324] env[65522]: DEBUG oslo_concurrency.lockutils [None req-94cb2c42-2e1e-4f8f-977e-b949dabd2e87 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Lock "bef9977d-ae21-426e-97d3-7b70a7276356" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.095s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1297.220136] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.220531] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.220822] env[65522]: DEBUG nova.objects.instance [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'resources' on Instance uuid 3adfefa5-fb19-44b2-b3c5-42f2e2918673 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1297.410609] env[65522]: INFO nova.compute.manager [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Rescuing [ 1297.410881] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquiring lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1297.411049] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquired lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1297.411217] env[65522]: DEBUG nova.network.neutron [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1297.594240] env[65522]: DEBUG oslo_concurrency.lockutils [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "7617189c-a902-42e7-8165-0e7c4a1de06d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.594528] env[65522]: DEBUG oslo_concurrency.lockutils [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "7617189c-a902-42e7-8165-0e7c4a1de06d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.594762] env[65522]: DEBUG oslo_concurrency.lockutils [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "7617189c-a902-42e7-8165-0e7c4a1de06d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.594946] env[65522]: DEBUG oslo_concurrency.lockutils [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "7617189c-a902-42e7-8165-0e7c4a1de06d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.595129] env[65522]: DEBUG oslo_concurrency.lockutils [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "7617189c-a902-42e7-8165-0e7c4a1de06d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1297.597335] env[65522]: INFO nova.compute.manager [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Terminating instance [ 1297.778115] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f344a8cf-917e-48cb-9e90-df5d583516ef {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.786323] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a1d0c8b-a099-464a-9fa3-173f6a953aa1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.817953] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e3a3164-7fb9-49ad-b5ef-970b9dd982bd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.825807] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5810ef31-be98-4f9d-8044-12638037cc7e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.841585] env[65522]: DEBUG nova.compute.provider_tree [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1297.914160] env[65522]: WARNING neutronclient.v2_0.client [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1297.914873] env[65522]: WARNING openstack [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1297.915225] env[65522]: WARNING openstack [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1298.032319] env[65522]: WARNING openstack [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1298.032717] env[65522]: WARNING openstack [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1298.092148] env[65522]: WARNING neutronclient.v2_0.client [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1298.092876] env[65522]: WARNING openstack [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1298.093263] env[65522]: WARNING openstack [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1298.101796] env[65522]: DEBUG nova.compute.manager [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1298.102014] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1298.103109] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c229d54-d537-48f7-ae00-777e47e12211 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.111229] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1298.111476] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cea24cd5-1854-4998-bd33-4dc5fccb57b7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.118380] env[65522]: DEBUG oslo_vmware.api [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1298.118380] env[65522]: value = "task-5115188" [ 1298.118380] env[65522]: _type = "Task" [ 1298.118380] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1298.128375] env[65522]: DEBUG oslo_vmware.api [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115188, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.185183] env[65522]: DEBUG nova.network.neutron [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Updating instance_info_cache with network_info: [{"id": "3a8642fb-edda-4cd0-8293-86da0093bee5", "address": "fa:16:3e:37:b0:01", "network": {"id": "6a449873-f579-41da-b6ae-a83934c5685a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1387860566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d30a4e7fde4842489a863e9a2c4cf8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "81d39ad2-4e62-4f09-a567-88ac5aa70467", "external-id": "nsx-vlan-transportzone-221", "segmentation_id": 221, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8642fb-ed", "ovs_interfaceid": "3a8642fb-edda-4cd0-8293-86da0093bee5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1298.344705] env[65522]: DEBUG nova.scheduler.client.report [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1298.629601] env[65522]: DEBUG oslo_vmware.api [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115188, 'name': PowerOffVM_Task, 'duration_secs': 0.232473} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1298.629874] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1298.630057] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1298.630324] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-54c9c615-23dc-4628-b215-078b438e2134 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.687877] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Releasing lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1298.691866] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1298.692102] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1298.692284] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleting the datastore file [datastore1] 7617189c-a902-42e7-8165-0e7c4a1de06d {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1298.692606] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6c5d0a93-8f40-45c7-87ba-b1d56d5c43e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.701047] env[65522]: DEBUG oslo_vmware.api [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for the task: (returnval){ [ 1298.701047] env[65522]: value = "task-5115190" [ 1298.701047] env[65522]: _type = "Task" [ 1298.701047] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1298.711601] env[65522]: DEBUG oslo_vmware.api [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115190, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.849863] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.629s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1298.875203] env[65522]: INFO nova.scheduler.client.report [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleted allocations for instance 3adfefa5-fb19-44b2-b3c5-42f2e2918673 [ 1299.211692] env[65522]: DEBUG oslo_vmware.api [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115190, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1299.384177] env[65522]: DEBUG oslo_concurrency.lockutils [None req-34fe9b4c-db17-4fa9-9f0d-308dc0ab9ac6 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "3adfefa5-fb19-44b2-b3c5-42f2e2918673" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.455s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1299.713044] env[65522]: DEBUG oslo_vmware.api [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Task: {'id': task-5115190, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.71421} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1299.713044] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1299.713044] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1299.713044] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1299.713044] env[65522]: INFO nova.compute.manager [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Took 1.61 seconds to destroy the instance on the hypervisor. [ 1299.713387] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1299.713521] env[65522]: DEBUG nova.compute.manager [-] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1299.713620] env[65522]: DEBUG nova.network.neutron [-] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1299.713857] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1299.714393] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1299.714684] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1299.751446] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1300.020476] env[65522]: DEBUG nova.compute.manager [req-022dff70-db07-4517-8ac3-a700d94a1ea3 req-c56eb4c9-81a1-4e19-9162-cc0bc13c8925 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Received event network-vif-deleted-dc729a30-fef3-4b2e-ab41-6c7a4eb89f73 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1300.020763] env[65522]: INFO nova.compute.manager [req-022dff70-db07-4517-8ac3-a700d94a1ea3 req-c56eb4c9-81a1-4e19-9162-cc0bc13c8925 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Neutron deleted interface dc729a30-fef3-4b2e-ab41-6c7a4eb89f73; detaching it from the instance and deleting it from the info cache [ 1300.020763] env[65522]: DEBUG nova.network.neutron [req-022dff70-db07-4517-8ac3-a700d94a1ea3 req-c56eb4c9-81a1-4e19-9162-cc0bc13c8925 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1300.225970] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1300.226291] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-961d3875-38b2-450f-a267-fd7722917969 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.234762] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1300.234762] env[65522]: value = "task-5115191" [ 1300.234762] env[65522]: _type = "Task" [ 1300.234762] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.244568] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115191, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.422354] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "084efb1b-dded-490b-a3fb-150417fe7588" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1300.422637] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "084efb1b-dded-490b-a3fb-150417fe7588" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1300.497984] env[65522]: DEBUG nova.network.neutron [-] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1300.523654] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-82e2e9c5-02d6-484a-82c8-b17f62e694e2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.533700] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931cce5f-801e-4b01-b54c-92ae03c91d07 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.561680] env[65522]: DEBUG nova.compute.manager [req-022dff70-db07-4517-8ac3-a700d94a1ea3 req-c56eb4c9-81a1-4e19-9162-cc0bc13c8925 service nova] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Detach interface failed, port_id=dc729a30-fef3-4b2e-ab41-6c7a4eb89f73, reason: Instance 7617189c-a902-42e7-8165-0e7c4a1de06d could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1300.745819] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115191, 'name': PowerOffVM_Task, 'duration_secs': 0.218238} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1300.746107] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1300.746972] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2369934e-c785-41de-9160-d2a61480bcf9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.766762] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c3ed56-32cd-42fa-b0c9-902707f4c768 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.799371] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1300.799687] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3c1df8c4-2456-4705-8779-28297bf8cdd3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.807601] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1300.807601] env[65522]: value = "task-5115192" [ 1300.807601] env[65522]: _type = "Task" [ 1300.807601] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.815633] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115192, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.925477] env[65522]: DEBUG nova.compute.manager [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1301.000794] env[65522]: INFO nova.compute.manager [-] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Took 1.29 seconds to deallocate network for instance. [ 1301.318234] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1301.318607] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1301.318650] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1301.318778] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1301.318956] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1301.319242] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2be4e15e-9113-4e70-8985-824c48594fe7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.328999] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1301.329214] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1301.329973] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6532064e-70cc-4ed9-8976-83dff9ef471a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.335769] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1301.335769] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526fc632-3603-9c02-2bf4-9a11b00847c5" [ 1301.335769] env[65522]: _type = "Task" [ 1301.335769] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1301.344509] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526fc632-3603-9c02-2bf4-9a11b00847c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.447696] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1301.448520] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1301.450038] env[65522]: INFO nova.compute.claims [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1301.507231] env[65522]: DEBUG oslo_concurrency.lockutils [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1301.846409] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]526fc632-3603-9c02-2bf4-9a11b00847c5, 'name': SearchDatastore_Task, 'duration_secs': 0.009825} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1301.847369] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe6518d6-9602-4990-8c60-7f21713e630e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.853504] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1301.853504] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522a9da6-3be9-11f8-fba4-fd94f071e564" [ 1301.853504] env[65522]: _type = "Task" [ 1301.853504] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1301.862015] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522a9da6-3be9-11f8-fba4-fd94f071e564, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.366018] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522a9da6-3be9-11f8-fba4-fd94f071e564, 'name': SearchDatastore_Task, 'duration_secs': 0.009624} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1302.366423] env[65522]: DEBUG oslo_concurrency.lockutils [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1302.366550] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] bef9977d-ae21-426e-97d3-7b70a7276356/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk. {{(pid=65522) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1302.366859] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d64cd25-aa58-49d2-84af-fc979f3e8401 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.374875] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1302.374875] env[65522]: value = "task-5115193" [ 1302.374875] env[65522]: _type = "Task" [ 1302.374875] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.383569] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115193, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.512215] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d81492-032e-41cd-ab76-7f53a8242c71 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.522588] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95bea2fb-f598-4e3d-aeb7-c497af9d745c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.559567] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74336f7-27b2-4d86-b1b0-a548034e0e3b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.569935] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9245f73-7228-4087-af13-a1f2a91f94fc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.587231] env[65522]: DEBUG nova.compute.provider_tree [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1302.886287] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115193, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471827} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1302.886523] env[65522]: INFO nova.virt.vmwareapi.ds_util [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore2] bef9977d-ae21-426e-97d3-7b70a7276356/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk. [ 1302.887371] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2deb2dda-a5a7-441d-974a-5dd845a7add0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.912833] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] bef9977d-ae21-426e-97d3-7b70a7276356/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1302.913247] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b22c315-6b49-457c-a2bb-6bd115c63a75 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.932175] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1302.932175] env[65522]: value = "task-5115194" [ 1302.932175] env[65522]: _type = "Task" [ 1302.932175] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.940391] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115194, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.090478] env[65522]: DEBUG nova.scheduler.client.report [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1303.444075] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115194, 'name': ReconfigVM_Task, 'duration_secs': 0.291903} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.444428] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Reconfigured VM instance instance-00000078 to attach disk [datastore2] bef9977d-ae21-426e-97d3-7b70a7276356/a889db67-7337-4e32-8e34-642f34402926-rescue.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1303.445129] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04897c0-c720-4bf0-9955-3c968fe05ebb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.470481] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32277ddb-182a-40f4-84b2-43d279e24c11 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.487413] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1303.487413] env[65522]: value = "task-5115195" [ 1303.487413] env[65522]: _type = "Task" [ 1303.487413] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.496467] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115195, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.595794] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.148s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1303.596362] env[65522]: DEBUG nova.compute.manager [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1303.599342] env[65522]: DEBUG oslo_concurrency.lockutils [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.092s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1303.599586] env[65522]: DEBUG nova.objects.instance [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lazy-loading 'resources' on Instance uuid 7617189c-a902-42e7-8165-0e7c4a1de06d {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1304.004011] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115195, 'name': ReconfigVM_Task, 'duration_secs': 0.166057} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.004312] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1304.004618] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a49e5eaf-e591-4c3b-8319-832fc55c29fa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.011976] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1304.011976] env[65522]: value = "task-5115196" [ 1304.011976] env[65522]: _type = "Task" [ 1304.011976] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.020968] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115196, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.102841] env[65522]: DEBUG nova.compute.utils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1304.104466] env[65522]: DEBUG nova.compute.manager [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1304.104690] env[65522]: DEBUG nova.network.neutron [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1304.105060] env[65522]: WARNING neutronclient.v2_0.client [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1304.105391] env[65522]: WARNING neutronclient.v2_0.client [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1304.105976] env[65522]: WARNING openstack [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1304.106331] env[65522]: WARNING openstack [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1304.161272] env[65522]: DEBUG nova.policy [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b51f1b47ad744dc2b2b0598cee11bc6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d1239b79ae94cceb89ae7a8bd57da08', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1304.164563] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39834556-4a88-4f03-b58e-715cfeabd3ba {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.172996] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26271302-2746-48ed-b33b-12a9923fe842 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.207863] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea851d1-3b7a-46d6-acb5-c9d6134adac9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.216417] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd0c685d-d272-44fe-b2b1-82aacb172e7c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.230784] env[65522]: DEBUG nova.compute.provider_tree [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1304.481766] env[65522]: DEBUG nova.network.neutron [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Successfully created port: 6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1304.523745] env[65522]: DEBUG oslo_vmware.api [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115196, 'name': PowerOnVM_Task, 'duration_secs': 0.475448} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.523994] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1304.526881] env[65522]: DEBUG nova.compute.manager [None req-55a71528-edd7-4603-922e-ff36fcadc74a tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1304.527985] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2d1b68-8059-4e36-9344-6ca929d3e2c3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.616534] env[65522]: DEBUG nova.compute.manager [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1304.734420] env[65522]: DEBUG nova.scheduler.client.report [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1305.238960] env[65522]: DEBUG oslo_concurrency.lockutils [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.639s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1305.266705] env[65522]: INFO nova.scheduler.client.report [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Deleted allocations for instance 7617189c-a902-42e7-8165-0e7c4a1de06d [ 1305.326950] env[65522]: DEBUG nova.compute.manager [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Received event network-changed-3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1305.326950] env[65522]: DEBUG nova.compute.manager [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Refreshing instance network info cache due to event network-changed-3a8642fb-edda-4cd0-8293-86da0093bee5. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1305.326950] env[65522]: DEBUG oslo_concurrency.lockutils [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] Acquiring lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1305.326950] env[65522]: DEBUG oslo_concurrency.lockutils [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] Acquired lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1305.326950] env[65522]: DEBUG nova.network.neutron [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Refreshing network info cache for port 3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1305.626284] env[65522]: DEBUG nova.compute.manager [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1305.654550] env[65522]: DEBUG nova.virt.hardware [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1305.654857] env[65522]: DEBUG nova.virt.hardware [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1305.654978] env[65522]: DEBUG nova.virt.hardware [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1305.655171] env[65522]: DEBUG nova.virt.hardware [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1305.655313] env[65522]: DEBUG nova.virt.hardware [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1305.655554] env[65522]: DEBUG nova.virt.hardware [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1305.655681] env[65522]: DEBUG nova.virt.hardware [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1305.655841] env[65522]: DEBUG nova.virt.hardware [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1305.656012] env[65522]: DEBUG nova.virt.hardware [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1305.656184] env[65522]: DEBUG nova.virt.hardware [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1305.656355] env[65522]: DEBUG nova.virt.hardware [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1305.657268] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d00f679-1a02-41f7-8a2d-58c3f91a3a5f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.665388] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60171569-43c3-4e31-9856-e1a0a01a5431 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.775351] env[65522]: DEBUG oslo_concurrency.lockutils [None req-65af365a-d8f5-42ca-8b53-4cad415b4d92 tempest-ServerActionsTestOtherA-1521945027 tempest-ServerActionsTestOtherA-1521945027-project-member] Lock "7617189c-a902-42e7-8165-0e7c4a1de06d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.181s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1305.833044] env[65522]: WARNING neutronclient.v2_0.client [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1305.833780] env[65522]: WARNING openstack [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1305.834238] env[65522]: WARNING openstack [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1305.920100] env[65522]: DEBUG nova.compute.manager [req-2a819c44-19f4-4bf2-9a94-5668e4d10792 req-c2d7ae89-bf0f-4b0b-88d2-1112e3e13792 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Received event network-vif-plugged-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1305.920324] env[65522]: DEBUG oslo_concurrency.lockutils [req-2a819c44-19f4-4bf2-9a94-5668e4d10792 req-c2d7ae89-bf0f-4b0b-88d2-1112e3e13792 service nova] Acquiring lock "084efb1b-dded-490b-a3fb-150417fe7588-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1305.920532] env[65522]: DEBUG oslo_concurrency.lockutils [req-2a819c44-19f4-4bf2-9a94-5668e4d10792 req-c2d7ae89-bf0f-4b0b-88d2-1112e3e13792 service nova] Lock "084efb1b-dded-490b-a3fb-150417fe7588-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1305.920700] env[65522]: DEBUG oslo_concurrency.lockutils [req-2a819c44-19f4-4bf2-9a94-5668e4d10792 req-c2d7ae89-bf0f-4b0b-88d2-1112e3e13792 service nova] Lock "084efb1b-dded-490b-a3fb-150417fe7588-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1305.920868] env[65522]: DEBUG nova.compute.manager [req-2a819c44-19f4-4bf2-9a94-5668e4d10792 req-c2d7ae89-bf0f-4b0b-88d2-1112e3e13792 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] No waiting events found dispatching network-vif-plugged-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1305.921051] env[65522]: WARNING nova.compute.manager [req-2a819c44-19f4-4bf2-9a94-5668e4d10792 req-c2d7ae89-bf0f-4b0b-88d2-1112e3e13792 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Received unexpected event network-vif-plugged-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed for instance with vm_state building and task_state spawning. [ 1305.994640] env[65522]: DEBUG nova.network.neutron [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Successfully updated port: 6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1306.169424] env[65522]: WARNING openstack [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1306.170033] env[65522]: WARNING openstack [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1306.253705] env[65522]: WARNING neutronclient.v2_0.client [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1306.253763] env[65522]: WARNING openstack [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1306.254084] env[65522]: WARNING openstack [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1306.344930] env[65522]: DEBUG nova.network.neutron [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Updated VIF entry in instance network info cache for port 3a8642fb-edda-4cd0-8293-86da0093bee5. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1306.345328] env[65522]: DEBUG nova.network.neutron [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Updating instance_info_cache with network_info: [{"id": "3a8642fb-edda-4cd0-8293-86da0093bee5", "address": "fa:16:3e:37:b0:01", "network": {"id": "6a449873-f579-41da-b6ae-a83934c5685a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1387860566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d30a4e7fde4842489a863e9a2c4cf8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "81d39ad2-4e62-4f09-a567-88ac5aa70467", "external-id": "nsx-vlan-transportzone-221", "segmentation_id": 221, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8642fb-ed", "ovs_interfaceid": "3a8642fb-edda-4cd0-8293-86da0093bee5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1306.498408] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1306.498408] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1306.498408] env[65522]: DEBUG nova.network.neutron [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1306.848331] env[65522]: DEBUG oslo_concurrency.lockutils [req-648c640e-b68e-4fed-82da-af7d0b4b98dc req-a0d92d04-ec7b-4a1d-867c-93d5205b6220 service nova] Releasing lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1307.004083] env[65522]: WARNING openstack [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1307.004624] env[65522]: WARNING openstack [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1307.051245] env[65522]: DEBUG nova.network.neutron [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1307.098192] env[65522]: WARNING openstack [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1307.098967] env[65522]: WARNING openstack [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1307.181423] env[65522]: WARNING neutronclient.v2_0.client [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1307.182093] env[65522]: WARNING openstack [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1307.182465] env[65522]: WARNING openstack [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1307.266552] env[65522]: DEBUG nova.network.neutron [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Updating instance_info_cache with network_info: [{"id": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "address": "fa:16:3e:87:bc:d7", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6071459f-85", "ovs_interfaceid": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1307.377219] env[65522]: DEBUG nova.compute.manager [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Received event network-changed-3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1307.377219] env[65522]: DEBUG nova.compute.manager [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Refreshing instance network info cache due to event network-changed-3a8642fb-edda-4cd0-8293-86da0093bee5. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1307.377219] env[65522]: DEBUG oslo_concurrency.lockutils [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Acquiring lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1307.378535] env[65522]: DEBUG oslo_concurrency.lockutils [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Acquired lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1307.378535] env[65522]: DEBUG nova.network.neutron [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Refreshing network info cache for port 3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1307.775021] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1307.775021] env[65522]: DEBUG nova.compute.manager [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Instance network_info: |[{"id": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "address": "fa:16:3e:87:bc:d7", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6071459f-85", "ovs_interfaceid": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1307.775021] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:bc:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a91c3a96-63d0-407c-bcde-c3d5b58d9cb2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1307.782963] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1307.783274] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1307.783564] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c79c8dba-ae08-4cea-9293-0cfb0c834ccd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.807694] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1307.807694] env[65522]: value = "task-5115197" [ 1307.807694] env[65522]: _type = "Task" [ 1307.807694] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.815661] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115197, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.881763] env[65522]: WARNING neutronclient.v2_0.client [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1307.882585] env[65522]: WARNING openstack [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1307.883100] env[65522]: WARNING openstack [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1307.981577] env[65522]: DEBUG nova.compute.manager [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Received event network-changed-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1307.981577] env[65522]: DEBUG nova.compute.manager [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Refreshing instance network info cache due to event network-changed-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1307.981577] env[65522]: DEBUG oslo_concurrency.lockutils [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] Acquiring lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1307.982151] env[65522]: DEBUG oslo_concurrency.lockutils [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] Acquired lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1307.982151] env[65522]: DEBUG nova.network.neutron [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Refreshing network info cache for port 6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1308.054066] env[65522]: WARNING openstack [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1308.054463] env[65522]: WARNING openstack [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1308.214576] env[65522]: WARNING neutronclient.v2_0.client [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1308.214576] env[65522]: WARNING openstack [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1308.214576] env[65522]: WARNING openstack [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1308.323222] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115197, 'name': CreateVM_Task, 'duration_secs': 0.407392} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.323589] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1308.323886] env[65522]: WARNING neutronclient.v2_0.client [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1308.324259] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1308.324538] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1308.324996] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1308.325149] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49c6bba9-568c-4fb2-b5c1-82abd0e271b9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.330550] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1308.330550] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ce2232-482d-3ef2-1745-b3413339fe66" [ 1308.330550] env[65522]: _type = "Task" [ 1308.330550] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1308.340599] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ce2232-482d-3ef2-1745-b3413339fe66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.343193] env[65522]: DEBUG nova.network.neutron [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Updated VIF entry in instance network info cache for port 3a8642fb-edda-4cd0-8293-86da0093bee5. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1308.343193] env[65522]: DEBUG nova.network.neutron [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Updating instance_info_cache with network_info: [{"id": "3a8642fb-edda-4cd0-8293-86da0093bee5", "address": "fa:16:3e:37:b0:01", "network": {"id": "6a449873-f579-41da-b6ae-a83934c5685a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1387860566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d30a4e7fde4842489a863e9a2c4cf8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "81d39ad2-4e62-4f09-a567-88ac5aa70467", "external-id": "nsx-vlan-transportzone-221", "segmentation_id": 221, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8642fb-ed", "ovs_interfaceid": "3a8642fb-edda-4cd0-8293-86da0093bee5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1308.485180] env[65522]: WARNING neutronclient.v2_0.client [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1308.486070] env[65522]: WARNING openstack [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1308.486427] env[65522]: WARNING openstack [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1308.636673] env[65522]: WARNING openstack [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1308.637083] env[65522]: WARNING openstack [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1308.730061] env[65522]: WARNING neutronclient.v2_0.client [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1308.730734] env[65522]: WARNING openstack [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1308.731079] env[65522]: WARNING openstack [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1308.815130] env[65522]: DEBUG nova.network.neutron [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Updated VIF entry in instance network info cache for port 6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1308.816051] env[65522]: DEBUG nova.network.neutron [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Updating instance_info_cache with network_info: [{"id": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "address": "fa:16:3e:87:bc:d7", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6071459f-85", "ovs_interfaceid": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1308.841883] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ce2232-482d-3ef2-1745-b3413339fe66, 'name': SearchDatastore_Task, 'duration_secs': 0.033922} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.842257] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1308.842671] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1308.842728] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1308.842838] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1308.843019] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1308.843487] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-93fb1607-6249-47cc-8b90-080061ca6cc7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.846256] env[65522]: DEBUG oslo_concurrency.lockutils [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Releasing lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1308.846953] env[65522]: DEBUG nova.compute.manager [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Received event network-changed-3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1308.846953] env[65522]: DEBUG nova.compute.manager [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Refreshing instance network info cache due to event network-changed-3a8642fb-edda-4cd0-8293-86da0093bee5. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1308.846953] env[65522]: DEBUG oslo_concurrency.lockutils [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Acquiring lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1308.846953] env[65522]: DEBUG oslo_concurrency.lockutils [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Acquired lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1308.847220] env[65522]: DEBUG nova.network.neutron [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Refreshing network info cache for port 3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1308.857012] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1308.857223] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1308.858282] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-678778cc-fafb-4677-8669-55b6b1a18582 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.865412] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1308.865412] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52aaf6cd-5237-4e4c-ad9e-ba2411da11f5" [ 1308.865412] env[65522]: _type = "Task" [ 1308.865412] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1308.874427] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52aaf6cd-5237-4e4c-ad9e-ba2411da11f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1309.319803] env[65522]: DEBUG oslo_concurrency.lockutils [req-9bdbea0f-960a-4879-9216-5011ecf33006 req-8cc2f31c-5a9b-4fc8-b182-6cd6363207b4 service nova] Releasing lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1309.350565] env[65522]: WARNING neutronclient.v2_0.client [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1309.351324] env[65522]: WARNING openstack [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1309.351688] env[65522]: WARNING openstack [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1309.376540] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52aaf6cd-5237-4e4c-ad9e-ba2411da11f5, 'name': SearchDatastore_Task, 'duration_secs': 0.010528} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1309.377266] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37981567-7006-473f-b899-9765526861a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.383543] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1309.383543] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ba5246-1529-6208-93d4-c3a940931e61" [ 1309.383543] env[65522]: _type = "Task" [ 1309.383543] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1309.394976] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ba5246-1529-6208-93d4-c3a940931e61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1309.410635] env[65522]: DEBUG nova.compute.manager [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Received event network-changed-3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1309.410722] env[65522]: DEBUG nova.compute.manager [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Refreshing instance network info cache due to event network-changed-3a8642fb-edda-4cd0-8293-86da0093bee5. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1309.410929] env[65522]: DEBUG oslo_concurrency.lockutils [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] Acquiring lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1309.462250] env[65522]: WARNING openstack [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1309.463107] env[65522]: WARNING openstack [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1309.515579] env[65522]: WARNING neutronclient.v2_0.client [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1309.516331] env[65522]: WARNING openstack [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1309.516739] env[65522]: WARNING openstack [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1309.559831] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquiring lock "bef9977d-ae21-426e-97d3-7b70a7276356" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1309.560130] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Lock "bef9977d-ae21-426e-97d3-7b70a7276356" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1309.560336] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquiring lock "bef9977d-ae21-426e-97d3-7b70a7276356-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1309.560527] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Lock "bef9977d-ae21-426e-97d3-7b70a7276356-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1309.560786] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Lock "bef9977d-ae21-426e-97d3-7b70a7276356-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1309.563835] env[65522]: INFO nova.compute.manager [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Terminating instance [ 1309.613080] env[65522]: DEBUG nova.network.neutron [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Updated VIF entry in instance network info cache for port 3a8642fb-edda-4cd0-8293-86da0093bee5. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1309.613443] env[65522]: DEBUG nova.network.neutron [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Updating instance_info_cache with network_info: [{"id": "3a8642fb-edda-4cd0-8293-86da0093bee5", "address": "fa:16:3e:37:b0:01", "network": {"id": "6a449873-f579-41da-b6ae-a83934c5685a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1387860566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d30a4e7fde4842489a863e9a2c4cf8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "81d39ad2-4e62-4f09-a567-88ac5aa70467", "external-id": "nsx-vlan-transportzone-221", "segmentation_id": 221, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8642fb-ed", "ovs_interfaceid": "3a8642fb-edda-4cd0-8293-86da0093bee5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1309.894576] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52ba5246-1529-6208-93d4-c3a940931e61, 'name': SearchDatastore_Task, 'duration_secs': 0.014778} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1309.894863] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1309.895146] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 084efb1b-dded-490b-a3fb-150417fe7588/084efb1b-dded-490b-a3fb-150417fe7588.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1309.895489] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-160ce49c-03f6-4842-99ea-6a22858d563b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.903250] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1309.903250] env[65522]: value = "task-5115198" [ 1309.903250] env[65522]: _type = "Task" [ 1309.903250] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1309.912209] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115198, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.068837] env[65522]: DEBUG nova.compute.manager [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1310.069153] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1310.070127] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7564dac1-f0c9-4dc4-9806-090690636d05 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.078780] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1310.079074] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-671195ef-cf8c-4272-860a-46bdeac85631 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.087504] env[65522]: DEBUG oslo_vmware.api [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1310.087504] env[65522]: value = "task-5115199" [ 1310.087504] env[65522]: _type = "Task" [ 1310.087504] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1310.096444] env[65522]: DEBUG oslo_vmware.api [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115199, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.116636] env[65522]: DEBUG oslo_concurrency.lockutils [req-8b5bbf8d-17fb-41ec-8e03-3071b5ec8dc0 req-a1c634b8-22b5-4a31-a802-3761b7f4be08 service nova] Releasing lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1310.117359] env[65522]: DEBUG oslo_concurrency.lockutils [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] Acquired lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1310.117657] env[65522]: DEBUG nova.network.neutron [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Refreshing network info cache for port 3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1310.418054] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115198, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.603877] env[65522]: DEBUG oslo_vmware.api [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115199, 'name': PowerOffVM_Task, 'duration_secs': 0.229025} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1310.604229] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1310.604481] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1310.604721] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5c28bce8-ecff-43bd-891c-746031c2af50 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.620574] env[65522]: WARNING neutronclient.v2_0.client [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1310.621296] env[65522]: WARNING openstack [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1310.621663] env[65522]: WARNING openstack [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1310.671355] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1310.671355] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Deleting contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1310.671555] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Deleting the datastore file [datastore2] bef9977d-ae21-426e-97d3-7b70a7276356 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1310.671796] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-16031a9a-0c99-4b1c-bc27-0a2a35369f51 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.680096] env[65522]: DEBUG oslo_vmware.api [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for the task: (returnval){ [ 1310.680096] env[65522]: value = "task-5115201" [ 1310.680096] env[65522]: _type = "Task" [ 1310.680096] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1310.689977] env[65522]: DEBUG oslo_vmware.api [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115201, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.727241] env[65522]: WARNING openstack [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1310.727856] env[65522]: WARNING openstack [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1310.777862] env[65522]: WARNING neutronclient.v2_0.client [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1310.778543] env[65522]: WARNING openstack [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1310.778897] env[65522]: WARNING openstack [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1310.869925] env[65522]: DEBUG nova.network.neutron [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Updated VIF entry in instance network info cache for port 3a8642fb-edda-4cd0-8293-86da0093bee5. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1310.870359] env[65522]: DEBUG nova.network.neutron [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Updating instance_info_cache with network_info: [{"id": "3a8642fb-edda-4cd0-8293-86da0093bee5", "address": "fa:16:3e:37:b0:01", "network": {"id": "6a449873-f579-41da-b6ae-a83934c5685a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1387860566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d30a4e7fde4842489a863e9a2c4cf8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "81d39ad2-4e62-4f09-a567-88ac5aa70467", "external-id": "nsx-vlan-transportzone-221", "segmentation_id": 221, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8642fb-ed", "ovs_interfaceid": "3a8642fb-edda-4cd0-8293-86da0093bee5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1310.917025] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115198, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.676297} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1310.917025] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 084efb1b-dded-490b-a3fb-150417fe7588/084efb1b-dded-490b-a3fb-150417fe7588.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1310.917025] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1310.917025] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03a83138-0c4c-49e8-9bb6-fbf31a3157cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.924238] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1310.924238] env[65522]: value = "task-5115202" [ 1310.924238] env[65522]: _type = "Task" [ 1310.924238] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1310.934773] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115202, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.191320] env[65522]: DEBUG oslo_vmware.api [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Task: {'id': task-5115201, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20764} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.191593] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1311.191793] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Deleted contents of the VM from datastore datastore2 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1311.191983] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1311.192181] env[65522]: INFO nova.compute.manager [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1311.193030] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1311.193030] env[65522]: DEBUG nova.compute.manager [-] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1311.193030] env[65522]: DEBUG nova.network.neutron [-] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1311.193030] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1311.194110] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1311.194110] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1311.246403] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1311.373396] env[65522]: DEBUG oslo_concurrency.lockutils [req-25f315a0-0d2a-4ba9-9186-232c41cfa39c req-d45c5159-8b8d-4f51-a15c-e14d3959df51 service nova] Releasing lock "refresh_cache-bef9977d-ae21-426e-97d3-7b70a7276356" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1311.437818] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115202, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073094} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.437818] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1311.439303] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ad61e3-56b5-4298-9259-d9efed2d2caa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.463388] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Reconfiguring VM instance instance-00000079 to attach disk [datastore1] 084efb1b-dded-490b-a3fb-150417fe7588/084efb1b-dded-490b-a3fb-150417fe7588.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1311.463745] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-114e7a16-0e3d-4119-bc8e-443aa5c3d25a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.486628] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1311.486628] env[65522]: value = "task-5115203" [ 1311.486628] env[65522]: _type = "Task" [ 1311.486628] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.501124] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115203, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.639642] env[65522]: DEBUG nova.compute.manager [req-97cd3a08-7e2f-4b46-885c-60525ff9b606 req-109e9f49-d5b9-4ffb-9d52-5280792cf778 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Received event network-vif-deleted-3a8642fb-edda-4cd0-8293-86da0093bee5 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1311.639767] env[65522]: INFO nova.compute.manager [req-97cd3a08-7e2f-4b46-885c-60525ff9b606 req-109e9f49-d5b9-4ffb-9d52-5280792cf778 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Neutron deleted interface 3a8642fb-edda-4cd0-8293-86da0093bee5; detaching it from the instance and deleting it from the info cache [ 1311.639958] env[65522]: DEBUG nova.network.neutron [req-97cd3a08-7e2f-4b46-885c-60525ff9b606 req-109e9f49-d5b9-4ffb-9d52-5280792cf778 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1311.997420] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115203, 'name': ReconfigVM_Task, 'duration_secs': 0.325321} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.997706] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Reconfigured VM instance instance-00000079 to attach disk [datastore1] 084efb1b-dded-490b-a3fb-150417fe7588/084efb1b-dded-490b-a3fb-150417fe7588.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1311.998409] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-64e19aaf-135d-4fec-865c-721c7855df1a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.002497] env[65522]: DEBUG nova.network.neutron [-] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1312.005719] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1312.005719] env[65522]: value = "task-5115204" [ 1312.005719] env[65522]: _type = "Task" [ 1312.005719] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.015962] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115204, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.145404] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bdb812c0-9a20-40b4-b6de-d572baae602e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.156866] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a802105b-1e8f-4ddb-b6f7-afe7235af892 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.185132] env[65522]: DEBUG nova.compute.manager [req-97cd3a08-7e2f-4b46-885c-60525ff9b606 req-109e9f49-d5b9-4ffb-9d52-5280792cf778 service nova] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Detach interface failed, port_id=3a8642fb-edda-4cd0-8293-86da0093bee5, reason: Instance bef9977d-ae21-426e-97d3-7b70a7276356 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1312.238205] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Acquiring lock "b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1312.238473] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Lock "b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1312.507117] env[65522]: INFO nova.compute.manager [-] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Took 1.31 seconds to deallocate network for instance. [ 1312.520794] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115204, 'name': Rename_Task, 'duration_secs': 0.147539} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.521069] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1312.521365] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06d12673-f9be-4016-b2c7-579002e3dc78 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.529012] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1312.529012] env[65522]: value = "task-5115205" [ 1312.529012] env[65522]: _type = "Task" [ 1312.529012] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.544079] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115205, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.741156] env[65522]: DEBUG nova.compute.manager [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1313.017190] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1313.017509] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1313.017786] env[65522]: DEBUG nova.objects.instance [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Lazy-loading 'resources' on Instance uuid bef9977d-ae21-426e-97d3-7b70a7276356 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1313.038866] env[65522]: DEBUG oslo_vmware.api [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115205, 'name': PowerOnVM_Task, 'duration_secs': 0.482696} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1313.039140] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1313.039336] env[65522]: INFO nova.compute.manager [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Took 7.41 seconds to spawn the instance on the hypervisor. [ 1313.039510] env[65522]: DEBUG nova.compute.manager [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1313.040296] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b4d03c-551d-4614-b670-03e6f6bc3b52 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.266757] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1313.559688] env[65522]: INFO nova.compute.manager [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Took 12.13 seconds to build instance. [ 1313.585400] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c6b214-096d-4529-ae28-27f55674568e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.599703] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780589ab-8e72-445f-9739-0bc43105b61b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.652911] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b150fdea-4dd9-4451-84e0-29dbda52b62e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.664043] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3168baac-d932-4265-b15e-92dfdbd6900d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.686897] env[65522]: DEBUG nova.compute.provider_tree [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1314.062220] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0fbff281-93b6-4473-b4ee-cd130a9dda2a tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "084efb1b-dded-490b-a3fb-150417fe7588" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.639s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1314.191195] env[65522]: DEBUG nova.scheduler.client.report [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1314.527446] env[65522]: DEBUG nova.compute.manager [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Received event network-changed-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1314.527609] env[65522]: DEBUG nova.compute.manager [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Refreshing instance network info cache due to event network-changed-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1314.527811] env[65522]: DEBUG oslo_concurrency.lockutils [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] Acquiring lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1314.527991] env[65522]: DEBUG oslo_concurrency.lockutils [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] Acquired lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1314.528166] env[65522]: DEBUG nova.network.neutron [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Refreshing network info cache for port 6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1314.696188] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.679s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1314.698581] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.432s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1314.700066] env[65522]: INFO nova.compute.claims [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1314.722631] env[65522]: INFO nova.scheduler.client.report [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Deleted allocations for instance bef9977d-ae21-426e-97d3-7b70a7276356 [ 1315.031164] env[65522]: WARNING neutronclient.v2_0.client [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1315.031929] env[65522]: WARNING openstack [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1315.032301] env[65522]: WARNING openstack [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1315.146922] env[65522]: WARNING openstack [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1315.147279] env[65522]: WARNING openstack [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1315.208574] env[65522]: WARNING neutronclient.v2_0.client [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1315.209216] env[65522]: WARNING openstack [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1315.209583] env[65522]: WARNING openstack [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1315.231356] env[65522]: DEBUG oslo_concurrency.lockutils [None req-cfa9f79c-1078-4464-922d-fad871f3c2b0 tempest-ServerRescueTestJSONUnderV235-729115084 tempest-ServerRescueTestJSONUnderV235-729115084-project-member] Lock "bef9977d-ae21-426e-97d3-7b70a7276356" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.671s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1315.297207] env[65522]: DEBUG nova.network.neutron [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Updated VIF entry in instance network info cache for port 6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1315.297595] env[65522]: DEBUG nova.network.neutron [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Updating instance_info_cache with network_info: [{"id": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "address": "fa:16:3e:87:bc:d7", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6071459f-85", "ovs_interfaceid": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1315.757076] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b56098-53aa-4192-b872-c71a8b0314b5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.765916] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8441a24f-1229-4cff-9b2f-4547ec848168 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.800044] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc539940-aae4-4478-9893-c9549d250ea0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.803318] env[65522]: DEBUG oslo_concurrency.lockutils [req-e9985c94-3476-4db9-83b1-7fc91f32b02c req-a213783d-d583-42d9-adbc-4fb12dd3caf5 service nova] Releasing lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1315.809887] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a504e56d-2108-4e6b-936a-3e2c32814720 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.826180] env[65522]: DEBUG nova.compute.provider_tree [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1316.329519] env[65522]: DEBUG nova.scheduler.client.report [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1316.834490] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.136s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1316.835021] env[65522]: DEBUG nova.compute.manager [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1317.339941] env[65522]: DEBUG nova.compute.utils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1317.341461] env[65522]: DEBUG nova.compute.manager [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1317.341682] env[65522]: DEBUG nova.network.neutron [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1317.342022] env[65522]: WARNING neutronclient.v2_0.client [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1317.342333] env[65522]: WARNING neutronclient.v2_0.client [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1317.342906] env[65522]: WARNING openstack [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1317.343338] env[65522]: WARNING openstack [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1317.399288] env[65522]: DEBUG nova.policy [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '606f05e13b8b403a9247920509eda0f7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e39365a6ab245548e8e6ccb96295934', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1317.699011] env[65522]: DEBUG nova.network.neutron [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Successfully created port: 3f619433-f69e-4f28-a540-0d072ce05297 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1317.852454] env[65522]: DEBUG nova.compute.manager [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1318.863151] env[65522]: DEBUG nova.compute.manager [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1318.889393] env[65522]: DEBUG nova.virt.hardware [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1318.889648] env[65522]: DEBUG nova.virt.hardware [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1318.889802] env[65522]: DEBUG nova.virt.hardware [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1318.889980] env[65522]: DEBUG nova.virt.hardware [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1318.890138] env[65522]: DEBUG nova.virt.hardware [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1318.890287] env[65522]: DEBUG nova.virt.hardware [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1318.890529] env[65522]: DEBUG nova.virt.hardware [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1318.890696] env[65522]: DEBUG nova.virt.hardware [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1318.890857] env[65522]: DEBUG nova.virt.hardware [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1318.891029] env[65522]: DEBUG nova.virt.hardware [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1318.891209] env[65522]: DEBUG nova.virt.hardware [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1318.892151] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb544d49-c61a-43ae-b49f-9535267d0965 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.901148] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574f80d3-fd40-4dd4-8d4d-50a7952be288 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.117685] env[65522]: DEBUG nova.compute.manager [req-a9ee3883-ec8e-44c3-ad28-77e178a65bda req-770fd03b-ad13-4c9c-9e94-8b6865a2412f service nova] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Received event network-vif-plugged-3f619433-f69e-4f28-a540-0d072ce05297 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1319.117906] env[65522]: DEBUG oslo_concurrency.lockutils [req-a9ee3883-ec8e-44c3-ad28-77e178a65bda req-770fd03b-ad13-4c9c-9e94-8b6865a2412f service nova] Acquiring lock "b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1319.118291] env[65522]: DEBUG oslo_concurrency.lockutils [req-a9ee3883-ec8e-44c3-ad28-77e178a65bda req-770fd03b-ad13-4c9c-9e94-8b6865a2412f service nova] Lock "b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1319.118556] env[65522]: DEBUG oslo_concurrency.lockutils [req-a9ee3883-ec8e-44c3-ad28-77e178a65bda req-770fd03b-ad13-4c9c-9e94-8b6865a2412f service nova] Lock "b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1319.118728] env[65522]: DEBUG nova.compute.manager [req-a9ee3883-ec8e-44c3-ad28-77e178a65bda req-770fd03b-ad13-4c9c-9e94-8b6865a2412f service nova] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] No waiting events found dispatching network-vif-plugged-3f619433-f69e-4f28-a540-0d072ce05297 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1319.118892] env[65522]: WARNING nova.compute.manager [req-a9ee3883-ec8e-44c3-ad28-77e178a65bda req-770fd03b-ad13-4c9c-9e94-8b6865a2412f service nova] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Received unexpected event network-vif-plugged-3f619433-f69e-4f28-a540-0d072ce05297 for instance with vm_state building and task_state spawning. [ 1319.204736] env[65522]: DEBUG nova.network.neutron [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Successfully updated port: 3f619433-f69e-4f28-a540-0d072ce05297 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1319.709486] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Acquiring lock "refresh_cache-b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1319.709486] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Acquired lock "refresh_cache-b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1319.709486] env[65522]: DEBUG nova.network.neutron [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1320.212192] env[65522]: WARNING openstack [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1320.212697] env[65522]: WARNING openstack [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1320.250029] env[65522]: DEBUG nova.network.neutron [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1320.269467] env[65522]: WARNING openstack [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1320.269839] env[65522]: WARNING openstack [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1320.338728] env[65522]: WARNING neutronclient.v2_0.client [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1320.339272] env[65522]: WARNING openstack [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1320.339689] env[65522]: WARNING openstack [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1320.436527] env[65522]: DEBUG nova.network.neutron [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Updating instance_info_cache with network_info: [{"id": "3f619433-f69e-4f28-a540-0d072ce05297", "address": "fa:16:3e:ec:11:17", "network": {"id": "f31244b6-5900-4e05-b5cb-a39af7a0e761", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-944515400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e39365a6ab245548e8e6ccb96295934", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f619433-f6", "ovs_interfaceid": "3f619433-f69e-4f28-a540-0d072ce05297", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1320.939640] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Releasing lock "refresh_cache-b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1320.940023] env[65522]: DEBUG nova.compute.manager [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Instance network_info: |[{"id": "3f619433-f69e-4f28-a540-0d072ce05297", "address": "fa:16:3e:ec:11:17", "network": {"id": "f31244b6-5900-4e05-b5cb-a39af7a0e761", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-944515400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e39365a6ab245548e8e6ccb96295934", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f619433-f6", "ovs_interfaceid": "3f619433-f69e-4f28-a540-0d072ce05297", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1320.940836] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:11:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c20f5114-0866-45b3-9a7c-62f113ff83fa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3f619433-f69e-4f28-a540-0d072ce05297', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1320.948462] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Creating folder: Project (3e39365a6ab245548e8e6ccb96295934). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1320.948808] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dbdc9ae1-4362-461c-9dbf-036f1aba8092 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.960742] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Created folder: Project (3e39365a6ab245548e8e6ccb96295934) in parent group-v994660. [ 1320.960926] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Creating folder: Instances. Parent ref: group-v994987. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1320.961192] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7731fa8e-7850-4614-938e-dc0632edb706 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.971060] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Created folder: Instances in parent group-v994987. [ 1320.971292] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1320.971481] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1320.971688] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e1efe92-b2d4-473b-8862-e63ee3ea2be4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.990393] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1320.990393] env[65522]: value = "task-5115208" [ 1320.990393] env[65522]: _type = "Task" [ 1320.990393] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.998101] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115208, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.150431] env[65522]: DEBUG nova.compute.manager [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Received event network-changed-3f619433-f69e-4f28-a540-0d072ce05297 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1321.150668] env[65522]: DEBUG nova.compute.manager [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Refreshing instance network info cache due to event network-changed-3f619433-f69e-4f28-a540-0d072ce05297. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1321.151023] env[65522]: DEBUG oslo_concurrency.lockutils [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] Acquiring lock "refresh_cache-b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1321.151507] env[65522]: DEBUG oslo_concurrency.lockutils [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] Acquired lock "refresh_cache-b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1321.151585] env[65522]: DEBUG nova.network.neutron [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Refreshing network info cache for port 3f619433-f69e-4f28-a540-0d072ce05297 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1321.500292] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115208, 'name': CreateVM_Task, 'duration_secs': 0.319096} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.500651] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1321.500955] env[65522]: WARNING neutronclient.v2_0.client [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1321.501324] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1321.501474] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1321.501785] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1321.502068] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-339fb193-def5-4acb-9315-1d764677910b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.506676] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Waiting for the task: (returnval){ [ 1321.506676] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d0a6c5-ed8f-bcb8-6c58-0bcbaaec7121" [ 1321.506676] env[65522]: _type = "Task" [ 1321.506676] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.514603] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d0a6c5-ed8f-bcb8-6c58-0bcbaaec7121, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.654688] env[65522]: WARNING neutronclient.v2_0.client [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1321.655401] env[65522]: WARNING openstack [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1321.655747] env[65522]: WARNING openstack [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1321.747093] env[65522]: WARNING openstack [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1321.747488] env[65522]: WARNING openstack [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1321.804982] env[65522]: WARNING neutronclient.v2_0.client [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1321.805651] env[65522]: WARNING openstack [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1321.806008] env[65522]: WARNING openstack [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1321.887537] env[65522]: DEBUG nova.network.neutron [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Updated VIF entry in instance network info cache for port 3f619433-f69e-4f28-a540-0d072ce05297. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1321.887905] env[65522]: DEBUG nova.network.neutron [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Updating instance_info_cache with network_info: [{"id": "3f619433-f69e-4f28-a540-0d072ce05297", "address": "fa:16:3e:ec:11:17", "network": {"id": "f31244b6-5900-4e05-b5cb-a39af7a0e761", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-944515400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e39365a6ab245548e8e6ccb96295934", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f619433-f6", "ovs_interfaceid": "3f619433-f69e-4f28-a540-0d072ce05297", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1322.017477] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52d0a6c5-ed8f-bcb8-6c58-0bcbaaec7121, 'name': SearchDatastore_Task, 'duration_secs': 0.01161} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1322.017775] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1322.018246] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1322.018503] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1322.018757] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1322.019063] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1322.019463] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c122a6d5-b17f-4e19-88c6-0e6b4029eb97 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.029350] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1322.029558] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1322.030303] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58652866-e93d-4c9d-8c0a-f637f4765267 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.035648] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Waiting for the task: (returnval){ [ 1322.035648] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520523c7-f4dc-fbd4-08d0-8d67afc9af59" [ 1322.035648] env[65522]: _type = "Task" [ 1322.035648] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.044364] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520523c7-f4dc-fbd4-08d0-8d67afc9af59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.391141] env[65522]: DEBUG oslo_concurrency.lockutils [req-14b345df-5f7a-4f54-ae96-ea09f5ac0440 req-d7ea4b38-bdd9-4543-96a8-d2b4b6f5f9bf service nova] Releasing lock "refresh_cache-b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1322.547644] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]520523c7-f4dc-fbd4-08d0-8d67afc9af59, 'name': SearchDatastore_Task, 'duration_secs': 0.009783} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1322.548453] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c5ea4ea-c95e-4c7b-99c8-1d5445c931fb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.554140] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Waiting for the task: (returnval){ [ 1322.554140] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e440d4-6fc0-deeb-4677-33046ff5a82b" [ 1322.554140] env[65522]: _type = "Task" [ 1322.554140] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.562252] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e440d4-6fc0-deeb-4677-33046ff5a82b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.065267] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e440d4-6fc0-deeb-4677-33046ff5a82b, 'name': SearchDatastore_Task, 'duration_secs': 0.010328} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.065538] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1323.065813] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187/b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1323.066099] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-62626362-edd0-4abb-b786-d9563b6372f6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.073590] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Waiting for the task: (returnval){ [ 1323.073590] env[65522]: value = "task-5115209" [ 1323.073590] env[65522]: _type = "Task" [ 1323.073590] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1323.081807] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': task-5115209, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.583787] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': task-5115209, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.427935} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.584200] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187/b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1323.584281] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1323.584497] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea4b953d-3196-4759-a52e-46c0a091d894 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.592359] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Waiting for the task: (returnval){ [ 1323.592359] env[65522]: value = "task-5115210" [ 1323.592359] env[65522]: _type = "Task" [ 1323.592359] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1323.600208] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': task-5115210, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.107104] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': task-5115210, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070215} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1324.107315] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1324.108403] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad35bbe-971b-4b26-ac2a-a3f110198f66 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.132463] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Reconfiguring VM instance instance-0000007a to attach disk [datastore1] b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187/b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1324.132737] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ababd30-75a6-4805-a2b8-f3cd4d953f25 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.153419] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Waiting for the task: (returnval){ [ 1324.153419] env[65522]: value = "task-5115211" [ 1324.153419] env[65522]: _type = "Task" [ 1324.153419] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.161575] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': task-5115211, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.663709] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': task-5115211, 'name': ReconfigVM_Task, 'duration_secs': 0.343783} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1324.664176] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Reconfigured VM instance instance-0000007a to attach disk [datastore1] b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187/b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1324.664588] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ab2cca04-459f-4cb5-a0fd-2434f7387237 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.670992] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Waiting for the task: (returnval){ [ 1324.670992] env[65522]: value = "task-5115212" [ 1324.670992] env[65522]: _type = "Task" [ 1324.670992] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.679019] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': task-5115212, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.181301] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': task-5115212, 'name': Rename_Task, 'duration_secs': 0.143848} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.181568] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1325.181824] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-44c7229b-01f8-45b9-b75a-e071597ee00f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.188994] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Waiting for the task: (returnval){ [ 1325.188994] env[65522]: value = "task-5115213" [ 1325.188994] env[65522]: _type = "Task" [ 1325.188994] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.199053] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': task-5115213, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.699359] env[65522]: DEBUG oslo_vmware.api [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': task-5115213, 'name': PowerOnVM_Task, 'duration_secs': 0.468061} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.699772] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1325.699823] env[65522]: INFO nova.compute.manager [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Took 6.84 seconds to spawn the instance on the hypervisor. [ 1325.700026] env[65522]: DEBUG nova.compute.manager [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1325.700931] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab81094-16fc-4d0a-9750-4dba9b56e3f2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.219842] env[65522]: INFO nova.compute.manager [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Took 12.97 seconds to build instance. [ 1326.722418] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ef733af0-db49-437c-936a-79b61f6cde4c tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Lock "b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.484s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1326.989613] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Acquiring lock "b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1326.989886] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Lock "b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1326.990098] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Acquiring lock "b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1326.990284] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Lock "b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1326.990447] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Lock "b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1326.992979] env[65522]: INFO nova.compute.manager [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Terminating instance [ 1327.496436] env[65522]: DEBUG nova.compute.manager [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1327.496642] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1327.497582] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48369731-7024-4d02-a592-24e57fbaad0c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.505372] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1327.505625] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa73d32f-0c49-48f5-a67e-4cf4ca4786cd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.512000] env[65522]: DEBUG oslo_vmware.api [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Waiting for the task: (returnval){ [ 1327.512000] env[65522]: value = "task-5115214" [ 1327.512000] env[65522]: _type = "Task" [ 1327.512000] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1327.520771] env[65522]: DEBUG oslo_vmware.api [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': task-5115214, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.022391] env[65522]: DEBUG oslo_vmware.api [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': task-5115214, 'name': PowerOffVM_Task, 'duration_secs': 0.182241} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.022794] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1328.022841] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1328.023101] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-44cbd1ea-86cf-4574-967f-8c49c690aa32 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.082705] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1328.082936] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1328.083092] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Deleting the datastore file [datastore1] b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1328.083377] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ffb04cb8-3d32-4d04-b3f7-691cf43af78d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.091940] env[65522]: DEBUG oslo_vmware.api [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Waiting for the task: (returnval){ [ 1328.091940] env[65522]: value = "task-5115216" [ 1328.091940] env[65522]: _type = "Task" [ 1328.091940] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.100129] env[65522]: DEBUG oslo_vmware.api [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': task-5115216, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.602280] env[65522]: DEBUG oslo_vmware.api [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Task: {'id': task-5115216, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136008} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.602567] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1328.602742] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1328.602927] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1328.603113] env[65522]: INFO nova.compute.manager [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1328.603362] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1328.603564] env[65522]: DEBUG nova.compute.manager [-] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1328.603668] env[65522]: DEBUG nova.network.neutron [-] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1328.603906] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1328.604457] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1328.604709] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1328.642262] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1328.897172] env[65522]: DEBUG nova.compute.manager [req-72621b08-4e76-47d7-afef-e4b3e5a07795 req-4415c9db-beb6-4d62-87ec-e5d65cbd82f0 service nova] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Received event network-vif-deleted-3f619433-f69e-4f28-a540-0d072ce05297 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1328.897375] env[65522]: INFO nova.compute.manager [req-72621b08-4e76-47d7-afef-e4b3e5a07795 req-4415c9db-beb6-4d62-87ec-e5d65cbd82f0 service nova] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Neutron deleted interface 3f619433-f69e-4f28-a540-0d072ce05297; detaching it from the instance and deleting it from the info cache [ 1328.897543] env[65522]: DEBUG nova.network.neutron [req-72621b08-4e76-47d7-afef-e4b3e5a07795 req-4415c9db-beb6-4d62-87ec-e5d65cbd82f0 service nova] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1329.371282] env[65522]: DEBUG nova.network.neutron [-] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1329.400563] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67d36b43-663b-47d0-a124-e72e8105196c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.411960] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbfbc24-57e9-4c87-8d22-ce5332e10853 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.442891] env[65522]: DEBUG nova.compute.manager [req-72621b08-4e76-47d7-afef-e4b3e5a07795 req-4415c9db-beb6-4d62-87ec-e5d65cbd82f0 service nova] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Detach interface failed, port_id=3f619433-f69e-4f28-a540-0d072ce05297, reason: Instance b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1329.874273] env[65522]: INFO nova.compute.manager [-] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Took 1.27 seconds to deallocate network for instance. [ 1330.380297] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1330.380814] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1330.380996] env[65522]: DEBUG nova.objects.instance [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Lazy-loading 'resources' on Instance uuid b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1330.928580] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad8b559-af4b-4c59-af7e-0bdb9d1519ae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.936431] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31da5525-31f2-4839-970d-4baedac1ee42 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.967148] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1996cd00-037e-4232-b8ec-058f3077b4fa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.975585] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab90b2cc-ad19-4cb7-a229-1560c442c407 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.989150] env[65522]: DEBUG nova.compute.provider_tree [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1331.492692] env[65522]: DEBUG nova.scheduler.client.report [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1331.999241] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1332.017260] env[65522]: INFO nova.scheduler.client.report [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Deleted allocations for instance b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187 [ 1332.525616] env[65522]: DEBUG oslo_concurrency.lockutils [None req-ec43740a-1453-4aff-b093-8a453f8546fe tempest-ServerMetadataNegativeTestJSON-1199686480 tempest-ServerMetadataNegativeTestJSON-1199686480-project-member] Lock "b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.536s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1335.770583] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquiring lock "b56fc96d-bad2-4655-b77f-66674a0c3330" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1335.770952] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Lock "b56fc96d-bad2-4655-b77f-66674a0c3330" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1336.273437] env[65522]: DEBUG nova.compute.manager [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1336.795404] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1336.795697] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1336.797172] env[65522]: INFO nova.compute.claims [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1337.843313] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcf076a-6929-4e48-8aa2-8cfb0959d3c2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.850956] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2de4ec-55eb-43da-8590-1b77621bcbb8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.882356] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9997e664-af8f-4145-96c6-d91aceeb5d7b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.890146] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b25a1b97-79b0-4675-98ac-5f15f7e61914 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.903562] env[65522]: DEBUG nova.compute.provider_tree [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1338.407116] env[65522]: DEBUG nova.scheduler.client.report [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1338.913070] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.116s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1338.913070] env[65522]: DEBUG nova.compute.manager [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1339.418322] env[65522]: DEBUG nova.compute.utils [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1339.419734] env[65522]: DEBUG nova.compute.manager [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Not allocating networking since 'none' was specified. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 1339.921745] env[65522]: DEBUG nova.compute.manager [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1340.931158] env[65522]: DEBUG nova.compute.manager [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1340.959123] env[65522]: DEBUG nova.virt.hardware [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1340.959393] env[65522]: DEBUG nova.virt.hardware [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1340.959550] env[65522]: DEBUG nova.virt.hardware [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1340.959738] env[65522]: DEBUG nova.virt.hardware [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1340.959879] env[65522]: DEBUG nova.virt.hardware [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1340.960036] env[65522]: DEBUG nova.virt.hardware [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1340.960253] env[65522]: DEBUG nova.virt.hardware [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1340.960415] env[65522]: DEBUG nova.virt.hardware [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1340.960618] env[65522]: DEBUG nova.virt.hardware [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1340.960800] env[65522]: DEBUG nova.virt.hardware [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1340.961013] env[65522]: DEBUG nova.virt.hardware [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1340.961979] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b2f445-5c7c-423c-ae67-13c31339ecb6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.971367] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b55dbb-fd68-4575-9e61-d41e55162cd6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.986385] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1340.991864] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Creating folder: Project (7f3da49f86234047a06ce387356aeae4). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1340.992186] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6661d3ed-c69c-4dc6-b7a8-58a2566c0a63 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.001903] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Created folder: Project (7f3da49f86234047a06ce387356aeae4) in parent group-v994660. [ 1341.002104] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Creating folder: Instances. Parent ref: group-v994990. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1341.002345] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff348348-50c5-4716-820f-dccb15d22431 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.010962] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Created folder: Instances in parent group-v994990. [ 1341.011326] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1341.011483] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1341.011715] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b3c9640e-0228-45fc-b9a2-d967bf6bb986 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.032381] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1341.032381] env[65522]: value = "task-5115219" [ 1341.032381] env[65522]: _type = "Task" [ 1341.032381] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.043603] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115219, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.542855] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115219, 'name': CreateVM_Task, 'duration_secs': 0.268809} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.543158] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1341.543666] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1341.543878] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1341.544266] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1341.544540] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b94b14e4-520c-4557-b198-8f8ff4b48c1f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.549884] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1341.549884] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dc56f4-32ac-8590-3146-3df41447108b" [ 1341.549884] env[65522]: _type = "Task" [ 1341.549884] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.558269] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dc56f4-32ac-8590-3146-3df41447108b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.060734] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52dc56f4-32ac-8590-3146-3df41447108b, 'name': SearchDatastore_Task, 'duration_secs': 0.011594} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.061127] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1342.061357] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1342.061599] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1342.061746] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1342.061928] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1342.062267] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d45841ea-34dc-4e7b-82dd-db40d8c944a3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.072281] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1342.072496] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1342.073312] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac5b13b3-40f7-4f8c-bb87-88b6fe7b5cfb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.079852] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1342.079852] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5208b040-4029-db87-c8ad-01af0616f6c7" [ 1342.079852] env[65522]: _type = "Task" [ 1342.079852] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1342.088452] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5208b040-4029-db87-c8ad-01af0616f6c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.590926] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]5208b040-4029-db87-c8ad-01af0616f6c7, 'name': SearchDatastore_Task, 'duration_secs': 0.01043} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.591752] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d50f979d-e485-4638-ba2d-85153304ca7a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.598087] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1342.598087] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52884d39-20a6-85e6-9959-692b35b049d1" [ 1342.598087] env[65522]: _type = "Task" [ 1342.598087] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1342.606223] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52884d39-20a6-85e6-9959-692b35b049d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1343.109028] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52884d39-20a6-85e6-9959-692b35b049d1, 'name': SearchDatastore_Task, 'duration_secs': 0.01005} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1343.109465] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1343.109590] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] b56fc96d-bad2-4655-b77f-66674a0c3330/b56fc96d-bad2-4655-b77f-66674a0c3330.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1343.109799] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e0c1b74c-632d-4dc8-a028-efba95cbea96 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.118719] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1343.118719] env[65522]: value = "task-5115220" [ 1343.118719] env[65522]: _type = "Task" [ 1343.118719] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1343.127480] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115220, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1343.629405] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115220, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435165} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1343.629648] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] b56fc96d-bad2-4655-b77f-66674a0c3330/b56fc96d-bad2-4655-b77f-66674a0c3330.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1343.629858] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1343.630147] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07cce069-09df-4ca9-b204-e4213f5385f4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.637081] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1343.637081] env[65522]: value = "task-5115221" [ 1343.637081] env[65522]: _type = "Task" [ 1343.637081] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1343.645512] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115221, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1344.147359] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115221, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077284} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1344.147801] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1344.148424] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ce8df3-e358-4ed4-9d3c-ae45d460f402 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.168462] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Reconfiguring VM instance instance-0000007b to attach disk [datastore1] b56fc96d-bad2-4655-b77f-66674a0c3330/b56fc96d-bad2-4655-b77f-66674a0c3330.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1344.168728] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9700735a-cca5-4862-868e-c4d7fc166ee2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.189416] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1344.189416] env[65522]: value = "task-5115222" [ 1344.189416] env[65522]: _type = "Task" [ 1344.189416] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.199188] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115222, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1344.699847] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115222, 'name': ReconfigVM_Task, 'duration_secs': 0.278097} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1344.700139] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Reconfigured VM instance instance-0000007b to attach disk [datastore1] b56fc96d-bad2-4655-b77f-66674a0c3330/b56fc96d-bad2-4655-b77f-66674a0c3330.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1344.700741] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2a8bf022-ad54-488b-bffd-be4069895491 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.707978] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1344.707978] env[65522]: value = "task-5115223" [ 1344.707978] env[65522]: _type = "Task" [ 1344.707978] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.716733] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115223, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.218386] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115223, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.719475] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115223, 'name': Rename_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.220138] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115223, 'name': Rename_Task, 'duration_secs': 1.144578} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.220617] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1346.220809] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a983b1d-d836-499d-9cbf-4f5835a76264 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.228492] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1346.228492] env[65522]: value = "task-5115224" [ 1346.228492] env[65522]: _type = "Task" [ 1346.228492] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.236837] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115224, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.738919] env[65522]: DEBUG oslo_vmware.api [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115224, 'name': PowerOnVM_Task, 'duration_secs': 0.425086} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.739204] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1346.739441] env[65522]: INFO nova.compute.manager [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Took 5.81 seconds to spawn the instance on the hypervisor. [ 1346.739639] env[65522]: DEBUG nova.compute.manager [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1346.740525] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f3f144-d6ac-491c-9429-5b5194130dc9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.256545] env[65522]: INFO nova.compute.manager [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Took 10.48 seconds to build instance. [ 1347.621705] env[65522]: INFO nova.compute.manager [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Rebuilding instance [ 1347.666951] env[65522]: DEBUG nova.compute.manager [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1347.667883] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8a56b1-3d21-4445-b645-f820b6854f51 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.758241] env[65522]: DEBUG oslo_concurrency.lockutils [None req-a7fb6e2f-8c27-44b4-945e-8a17893c2855 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Lock "b56fc96d-bad2-4655-b77f-66674a0c3330" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.987s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1348.682561] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1348.683034] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88beddc4-aa9a-48ce-b196-cb80f434e224 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.690623] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1348.690623] env[65522]: value = "task-5115225" [ 1348.690623] env[65522]: _type = "Task" [ 1348.690623] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1348.700568] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115225, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.201412] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115225, 'name': PowerOffVM_Task, 'duration_secs': 0.193204} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1349.201743] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1349.202028] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1349.202851] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-237f8f1a-6f4e-4903-b305-6519c3aa7747 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.209759] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1349.209988] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fae8a268-eb35-4347-be77-2aae9c4c1a33 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.236446] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1349.236653] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1349.236836] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Deleting the datastore file [datastore1] b56fc96d-bad2-4655-b77f-66674a0c3330 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1349.237251] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7af87fc-bfe3-4e66-a360-f5632de4bcde {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.244763] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1349.244763] env[65522]: value = "task-5115227" [ 1349.244763] env[65522]: _type = "Task" [ 1349.244763] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1349.253464] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115227, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.758366] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115227, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102029} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1349.758861] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1349.759240] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1349.759584] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1350.791465] env[65522]: DEBUG nova.virt.hardware [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1350.791776] env[65522]: DEBUG nova.virt.hardware [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1350.791882] env[65522]: DEBUG nova.virt.hardware [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1350.792072] env[65522]: DEBUG nova.virt.hardware [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1350.792226] env[65522]: DEBUG nova.virt.hardware [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1350.792370] env[65522]: DEBUG nova.virt.hardware [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1350.792576] env[65522]: DEBUG nova.virt.hardware [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1350.792731] env[65522]: DEBUG nova.virt.hardware [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1350.792894] env[65522]: DEBUG nova.virt.hardware [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1350.793106] env[65522]: DEBUG nova.virt.hardware [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1350.793316] env[65522]: DEBUG nova.virt.hardware [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1350.794492] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0207c79-c5e2-4046-9205-6a8c0f830d35 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.803452] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872ac1d4-76a2-4a55-b54c-6a56646dfb0b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.818182] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Instance VIF info [] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1350.823722] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1350.823967] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1350.824206] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cf511d6a-60fd-41be-82e6-d906f62d4155 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.841353] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1350.841353] env[65522]: value = "task-5115228" [ 1350.841353] env[65522]: _type = "Task" [ 1350.841353] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1350.849420] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115228, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1351.351824] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115228, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1351.715563] env[65522]: DEBUG oslo_concurrency.lockutils [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "084efb1b-dded-490b-a3fb-150417fe7588" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1351.715848] env[65522]: DEBUG oslo_concurrency.lockutils [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "084efb1b-dded-490b-a3fb-150417fe7588" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1351.716047] env[65522]: INFO nova.compute.manager [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Shelving [ 1351.854942] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115228, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1352.019236] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1352.019597] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1352.019695] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1352.019842] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1352.019990] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1352.020179] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1352.020403] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1352.020565] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1352.020811] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1352.353521] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115228, 'name': CreateVM_Task, 'duration_secs': 1.267585} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1352.353734] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1352.354134] env[65522]: DEBUG oslo_concurrency.lockutils [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1352.354292] env[65522]: DEBUG oslo_concurrency.lockutils [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1352.354627] env[65522]: DEBUG oslo_concurrency.lockutils [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1352.354892] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6a87a66-7392-4a55-b4fc-6bbec633fedd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.360512] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1352.360512] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522e0772-676b-79ac-ee02-c83b3eeb171e" [ 1352.360512] env[65522]: _type = "Task" [ 1352.360512] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1352.370179] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522e0772-676b-79ac-ee02-c83b3eeb171e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1352.524568] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1352.524946] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1352.525250] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1352.525465] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1352.526732] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc57bdff-6ded-4cf5-8db2-f1d95ff49891 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.538856] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8eeb2d-75bd-4752-a87d-7cc030763bbd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.561879] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d686058d-3483-47e9-98e2-8394aaf00513 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.572414] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-599eb3ec-1b6d-49a4-8d11-4147263e5638 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.619445] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180106MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1352.619699] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1352.620083] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1352.724739] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1352.725112] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f226e632-82a5-4aba-8dab-6cc591ca206c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.732983] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1352.732983] env[65522]: value = "task-5115229" [ 1352.732983] env[65522]: _type = "Task" [ 1352.732983] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1352.742094] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115229, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1352.871726] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]522e0772-676b-79ac-ee02-c83b3eeb171e, 'name': SearchDatastore_Task, 'duration_secs': 0.010558} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1352.872139] env[65522]: DEBUG oslo_concurrency.lockutils [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1352.872323] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1352.872596] env[65522]: DEBUG oslo_concurrency.lockutils [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1352.872739] env[65522]: DEBUG oslo_concurrency.lockutils [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1352.872929] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1352.873255] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e791192-f18a-41e0-9269-e241d2e91381 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.883374] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1352.883548] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1352.884312] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc317bb2-1d2f-4de6-afab-15e2caa6a654 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.890449] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1352.890449] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cc9427-68d9-c73c-3dc5-50e50b2fd110" [ 1352.890449] env[65522]: _type = "Task" [ 1352.890449] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1352.901075] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cc9427-68d9-c73c-3dc5-50e50b2fd110, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1353.243136] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115229, 'name': PowerOffVM_Task, 'duration_secs': 0.254202} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1353.243403] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1353.244227] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6721b5-f705-4728-b8f0-747d61b21ee8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.263066] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f75ced-2e4a-4e18-9916-33632db81305 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.401649] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52cc9427-68d9-c73c-3dc5-50e50b2fd110, 'name': SearchDatastore_Task, 'duration_secs': 0.010398} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1353.402432] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d15bf6f1-f163-44af-97c0-0aaf34781466 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.409347] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1353.409347] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bcc38a-5c5c-956a-8778-510e0bf7f7f4" [ 1353.409347] env[65522]: _type = "Task" [ 1353.409347] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1353.418012] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bcc38a-5c5c-956a-8778-510e0bf7f7f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1353.647491] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 084efb1b-dded-490b-a3fb-150417fe7588 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1353.647491] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance b56fc96d-bad2-4655-b77f-66674a0c3330 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1353.647692] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1353.647887] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=100GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '2', 'num_vm_active': '2', 'num_task_shelving': '1', 'num_os_type_None': '2', 'num_proj_0d1239b79ae94cceb89ae7a8bd57da08': '1', 'io_workload': '0', 'num_task_rebuild_spawning': '1', 'num_proj_7f3da49f86234047a06ce387356aeae4': '1'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1353.694318] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8df88a-9f39-48ee-958a-27a9f2fb942c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.703030] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1ae9d7-7e81-4de4-a5bd-6fa8e6bd1cc5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.734643] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1d8a57-6763-4295-96ca-997f80d45ec9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.742605] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53659866-6124-4bf4-accc-d031b81f538d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.756015] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1353.773960] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Creating Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1353.774279] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8a6e8901-35a7-4e3f-8fc6-6e765380f0f8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.782380] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1353.782380] env[65522]: value = "task-5115230" [ 1353.782380] env[65522]: _type = "Task" [ 1353.782380] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1353.790814] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115230, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1353.919782] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52bcc38a-5c5c-956a-8778-510e0bf7f7f4, 'name': SearchDatastore_Task, 'duration_secs': 0.010966} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1353.920201] env[65522]: DEBUG oslo_concurrency.lockutils [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1353.920250] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] b56fc96d-bad2-4655-b77f-66674a0c3330/b56fc96d-bad2-4655-b77f-66674a0c3330.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1353.920516] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c108bf79-e7c2-49ce-b44f-dd157dd40b1d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.927091] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1353.927091] env[65522]: value = "task-5115231" [ 1353.927091] env[65522]: _type = "Task" [ 1353.927091] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1353.935641] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115231, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.260514] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1354.297336] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115230, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.437606] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115231, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445994} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1354.437898] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] b56fc96d-bad2-4655-b77f-66674a0c3330/b56fc96d-bad2-4655-b77f-66674a0c3330.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1354.438118] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1354.438375] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-327b47d4-7f7b-4564-8ef5-b170906bcc48 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.444774] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1354.444774] env[65522]: value = "task-5115232" [ 1354.444774] env[65522]: _type = "Task" [ 1354.444774] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.452455] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115232, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.768246] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1354.768442] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.148s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1354.797459] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115230, 'name': CreateSnapshot_Task, 'duration_secs': 0.52648} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1354.797899] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Created Snapshot of the VM instance {{(pid=65522) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1354.799157] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-298cea3c-234e-44a3-b7a3-9da34e3c5361 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.955290] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115232, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067544} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1354.955757] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1354.956362] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a0b27f9-2319-4b0a-a819-b08e4e5bc6fa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.976324] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Reconfiguring VM instance instance-0000007b to attach disk [datastore1] b56fc96d-bad2-4655-b77f-66674a0c3330/b56fc96d-bad2-4655-b77f-66674a0c3330.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1354.976913] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be34fc26-dc82-4e54-bbaf-79e3387ae0d5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.999335] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1354.999335] env[65522]: value = "task-5115233" [ 1354.999335] env[65522]: _type = "Task" [ 1354.999335] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1355.007403] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115233, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.323367] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Creating linked-clone VM from snapshot {{(pid=65522) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1355.324240] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-52a48cdc-da07-4ae4-b762-8c29cbc2e4a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.333997] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1355.333997] env[65522]: value = "task-5115234" [ 1355.333997] env[65522]: _type = "Task" [ 1355.333997] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1355.342526] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115234, 'name': CloneVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.510471] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115233, 'name': ReconfigVM_Task, 'duration_secs': 0.294446} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1355.510779] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Reconfigured VM instance instance-0000007b to attach disk [datastore1] b56fc96d-bad2-4655-b77f-66674a0c3330/b56fc96d-bad2-4655-b77f-66674a0c3330.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1355.511464] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1860ce2d-e80e-457f-b147-f5815050ef23 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.520818] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1355.520818] env[65522]: value = "task-5115235" [ 1355.520818] env[65522]: _type = "Task" [ 1355.520818] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1355.534307] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115235, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.844735] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115234, 'name': CloneVM_Task} progress is 94%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.032188] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115235, 'name': Rename_Task, 'duration_secs': 0.149317} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1356.032584] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1356.032923] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5bcb3454-00d1-45d3-9209-078153271ecd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.040581] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1356.040581] env[65522]: value = "task-5115236" [ 1356.040581] env[65522]: _type = "Task" [ 1356.040581] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1356.050620] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115236, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.345053] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115234, 'name': CloneVM_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.550913] env[65522]: DEBUG oslo_vmware.api [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115236, 'name': PowerOnVM_Task, 'duration_secs': 0.450295} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1356.551200] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1356.551400] env[65522]: DEBUG nova.compute.manager [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1356.552194] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73313cf4-2257-4449-853f-83bd5b99eecd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.846219] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115234, 'name': CloneVM_Task, 'duration_secs': 1.083353} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1356.846573] env[65522]: INFO nova.virt.vmwareapi.vmops [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Created linked-clone VM from snapshot [ 1356.847361] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c2be92-ca41-4329-895c-3d62bb82ef1a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.861031] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Uploading image bbfe6c9f-9556-4717-8ecd-51b706a27303 {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1356.888924] env[65522]: DEBUG oslo_vmware.rw_handles [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1356.888924] env[65522]: value = "vm-994995" [ 1356.888924] env[65522]: _type = "VirtualMachine" [ 1356.888924] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1356.890274] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-490b2845-1931-44d2-92f2-92170201c409 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.897046] env[65522]: DEBUG oslo_vmware.rw_handles [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lease: (returnval){ [ 1356.897046] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f53de5-af08-b607-a2f4-85c543dd7647" [ 1356.897046] env[65522]: _type = "HttpNfcLease" [ 1356.897046] env[65522]: } obtained for exporting VM: (result){ [ 1356.897046] env[65522]: value = "vm-994995" [ 1356.897046] env[65522]: _type = "VirtualMachine" [ 1356.897046] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1356.897428] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the lease: (returnval){ [ 1356.897428] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f53de5-af08-b607-a2f4-85c543dd7647" [ 1356.897428] env[65522]: _type = "HttpNfcLease" [ 1356.897428] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1356.905644] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1356.905644] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f53de5-af08-b607-a2f4-85c543dd7647" [ 1356.905644] env[65522]: _type = "HttpNfcLease" [ 1356.905644] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1357.068925] env[65522]: DEBUG oslo_concurrency.lockutils [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1357.069395] env[65522]: DEBUG oslo_concurrency.lockutils [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1357.069395] env[65522]: DEBUG nova.objects.instance [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65522) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1357.356941] env[65522]: DEBUG oslo_concurrency.lockutils [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquiring lock "b56fc96d-bad2-4655-b77f-66674a0c3330" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1357.357185] env[65522]: DEBUG oslo_concurrency.lockutils [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Lock "b56fc96d-bad2-4655-b77f-66674a0c3330" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1357.357440] env[65522]: DEBUG oslo_concurrency.lockutils [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquiring lock "b56fc96d-bad2-4655-b77f-66674a0c3330-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1357.357630] env[65522]: DEBUG oslo_concurrency.lockutils [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Lock "b56fc96d-bad2-4655-b77f-66674a0c3330-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1357.357799] env[65522]: DEBUG oslo_concurrency.lockutils [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Lock "b56fc96d-bad2-4655-b77f-66674a0c3330-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1357.360097] env[65522]: INFO nova.compute.manager [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Terminating instance [ 1357.405849] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1357.405849] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f53de5-af08-b607-a2f4-85c543dd7647" [ 1357.405849] env[65522]: _type = "HttpNfcLease" [ 1357.405849] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1357.406360] env[65522]: DEBUG oslo_vmware.rw_handles [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1357.406360] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52f53de5-af08-b607-a2f4-85c543dd7647" [ 1357.406360] env[65522]: _type = "HttpNfcLease" [ 1357.406360] env[65522]: }. {{(pid=65522) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1357.407332] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9315c41b-a578-469a-9b6f-f70ff396a6c9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.415261] env[65522]: DEBUG oslo_vmware.rw_handles [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e025b6-0dff-4b03-36d2-8e2472e432bb/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1357.415430] env[65522]: DEBUG oslo_vmware.rw_handles [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e025b6-0dff-4b03-36d2-8e2472e432bb/disk-0.vmdk for reading. {{(pid=65522) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1357.505092] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d6489d19-76db-409f-904f-f89ba4d4baf5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.863759] env[65522]: DEBUG oslo_concurrency.lockutils [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquiring lock "refresh_cache-b56fc96d-bad2-4655-b77f-66674a0c3330" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1357.863982] env[65522]: DEBUG oslo_concurrency.lockutils [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquired lock "refresh_cache-b56fc96d-bad2-4655-b77f-66674a0c3330" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1357.864130] env[65522]: DEBUG nova.network.neutron [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1358.079315] env[65522]: DEBUG oslo_concurrency.lockutils [None req-07b4ca6e-9e51-4525-9ac8-c4e1fb276f7f tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1358.367403] env[65522]: WARNING neutronclient.v2_0.client [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1358.368080] env[65522]: WARNING openstack [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1358.368431] env[65522]: WARNING openstack [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1358.391792] env[65522]: DEBUG nova.network.neutron [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1358.447865] env[65522]: DEBUG nova.network.neutron [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1358.951044] env[65522]: DEBUG oslo_concurrency.lockutils [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Releasing lock "refresh_cache-b56fc96d-bad2-4655-b77f-66674a0c3330" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1358.951592] env[65522]: DEBUG nova.compute.manager [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1358.951865] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1358.952895] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b90b38-28cc-487a-aa8c-ad7a76c2fc6b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.961684] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1358.962093] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-217876e4-3b71-4889-8215-f1c9501c9f78 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.969323] env[65522]: DEBUG oslo_vmware.api [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1358.969323] env[65522]: value = "task-5115238" [ 1358.969323] env[65522]: _type = "Task" [ 1358.969323] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1358.980581] env[65522]: DEBUG oslo_vmware.api [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115238, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.479754] env[65522]: DEBUG oslo_vmware.api [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115238, 'name': PowerOffVM_Task, 'duration_secs': 0.125131} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.480156] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1359.480333] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1359.480786] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-88c00802-ad8d-4c0c-b78b-5f8ea6f406f8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.506021] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1359.506666] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1359.506907] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Deleting the datastore file [datastore1] b56fc96d-bad2-4655-b77f-66674a0c3330 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1359.507242] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b1966481-fa3f-4300-bf51-4e5766084659 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.515842] env[65522]: DEBUG oslo_vmware.api [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for the task: (returnval){ [ 1359.515842] env[65522]: value = "task-5115240" [ 1359.515842] env[65522]: _type = "Task" [ 1359.515842] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.525314] env[65522]: DEBUG oslo_vmware.api [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115240, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.025477] env[65522]: DEBUG oslo_vmware.api [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Task: {'id': task-5115240, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.109001} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.025726] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1360.025908] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1360.026125] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1360.026329] env[65522]: INFO nova.compute.manager [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1360.026582] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1360.026773] env[65522]: DEBUG nova.compute.manager [-] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1360.026870] env[65522]: DEBUG nova.network.neutron [-] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1360.027134] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1360.027661] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1360.027908] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1360.048916] env[65522]: DEBUG nova.network.neutron [-] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1360.049191] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1360.551917] env[65522]: DEBUG nova.network.neutron [-] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1361.055273] env[65522]: INFO nova.compute.manager [-] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Took 1.03 seconds to deallocate network for instance. [ 1361.562419] env[65522]: DEBUG oslo_concurrency.lockutils [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1361.562904] env[65522]: DEBUG oslo_concurrency.lockutils [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1361.562904] env[65522]: DEBUG nova.objects.instance [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Lazy-loading 'resources' on Instance uuid b56fc96d-bad2-4655-b77f-66674a0c3330 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1362.108137] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743dc63a-1de8-4932-9b14-7eff6ea51d67 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.116291] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9883290-0555-466d-a97b-f03145213db9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.146622] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694bf8d0-12bf-41d7-a2a0-e51c79a85a7f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.154908] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c6ca54-16a0-4d34-8608-ba1e24009782 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.171044] env[65522]: DEBUG nova.compute.provider_tree [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1362.675364] env[65522]: DEBUG nova.scheduler.client.report [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1363.181137] env[65522]: DEBUG oslo_concurrency.lockutils [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1363.204788] env[65522]: INFO nova.scheduler.client.report [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Deleted allocations for instance b56fc96d-bad2-4655-b77f-66674a0c3330 [ 1363.716033] env[65522]: DEBUG oslo_concurrency.lockutils [None req-caf66bf0-4a36-40d5-a6c2-258cde5c1f00 tempest-ServerShowV254Test-512345036 tempest-ServerShowV254Test-512345036-project-member] Lock "b56fc96d-bad2-4655-b77f-66674a0c3330" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.358s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1365.889633] env[65522]: DEBUG oslo_vmware.rw_handles [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e025b6-0dff-4b03-36d2-8e2472e432bb/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1365.890686] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c96c6d4-fdcd-4e8c-9023-d581863b0837 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.897602] env[65522]: DEBUG oslo_vmware.rw_handles [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e025b6-0dff-4b03-36d2-8e2472e432bb/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1365.897756] env[65522]: ERROR oslo_vmware.rw_handles [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e025b6-0dff-4b03-36d2-8e2472e432bb/disk-0.vmdk due to incomplete transfer. [ 1365.897987] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2dc7997c-dcf0-4ea7-bb7a-8881a655147d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.907898] env[65522]: DEBUG oslo_vmware.rw_handles [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e025b6-0dff-4b03-36d2-8e2472e432bb/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1365.908121] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Uploaded image bbfe6c9f-9556-4717-8ecd-51b706a27303 to the Glance image server {{(pid=65522) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1365.910801] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Destroying the VM {{(pid=65522) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1365.911264] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b2e363c9-9042-412e-9c44-b2bdb8f712c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.918364] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1365.918364] env[65522]: value = "task-5115241" [ 1365.918364] env[65522]: _type = "Task" [ 1365.918364] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1365.926985] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115241, 'name': Destroy_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1366.428073] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115241, 'name': Destroy_Task, 'duration_secs': 0.366654} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1366.428382] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Destroyed the VM [ 1366.428622] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Deleting Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1366.428878] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-150332cf-f84a-4114-ae76-d8128daf1982 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.434876] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1366.434876] env[65522]: value = "task-5115242" [ 1366.434876] env[65522]: _type = "Task" [ 1366.434876] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1366.443416] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115242, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1366.945909] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115242, 'name': RemoveSnapshot_Task, 'duration_secs': 0.336843} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1366.945909] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Deleted Snapshot of the VM instance {{(pid=65522) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1366.946345] env[65522]: DEBUG nova.compute.manager [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1366.947049] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19562201-cd6d-4a51-bde4-c313682ef0a8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.459349] env[65522]: INFO nova.compute.manager [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Shelve offloading [ 1367.963144] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1367.963549] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-863ac42b-cfd4-4d75-926b-e9af38d6a6e7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.970991] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1367.970991] env[65522]: value = "task-5115243" [ 1367.970991] env[65522]: _type = "Task" [ 1367.970991] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1367.979501] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115243, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1368.481734] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] VM already powered off {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1368.481937] env[65522]: DEBUG nova.compute.manager [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1368.482734] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544683c6-1fc4-4d92-90e9-a085bd914768 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.488596] env[65522]: DEBUG oslo_concurrency.lockutils [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1368.488758] env[65522]: DEBUG oslo_concurrency.lockutils [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1368.488921] env[65522]: DEBUG nova.network.neutron [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1368.991891] env[65522]: WARNING neutronclient.v2_0.client [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1368.992648] env[65522]: WARNING openstack [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1368.993033] env[65522]: WARNING openstack [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1369.101376] env[65522]: WARNING openstack [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1369.101768] env[65522]: WARNING openstack [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1369.161283] env[65522]: WARNING neutronclient.v2_0.client [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1369.161926] env[65522]: WARNING openstack [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1369.162334] env[65522]: WARNING openstack [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1369.239443] env[65522]: DEBUG nova.network.neutron [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Updating instance_info_cache with network_info: [{"id": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "address": "fa:16:3e:87:bc:d7", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6071459f-85", "ovs_interfaceid": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1369.742550] env[65522]: DEBUG oslo_concurrency.lockutils [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1369.743046] env[65522]: WARNING neutronclient.v2_0.client [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1369.743691] env[65522]: WARNING openstack [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1369.744061] env[65522]: WARNING openstack [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1369.749177] env[65522]: WARNING neutronclient.v2_0.client [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1369.989489] env[65522]: DEBUG nova.compute.manager [req-87598f4a-96ba-4134-bbce-e53d042a294e req-31612da2-c92c-4cce-beca-382f73ae28c0 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Received event network-vif-unplugged-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1369.989716] env[65522]: DEBUG oslo_concurrency.lockutils [req-87598f4a-96ba-4134-bbce-e53d042a294e req-31612da2-c92c-4cce-beca-382f73ae28c0 service nova] Acquiring lock "084efb1b-dded-490b-a3fb-150417fe7588-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1369.989919] env[65522]: DEBUG oslo_concurrency.lockutils [req-87598f4a-96ba-4134-bbce-e53d042a294e req-31612da2-c92c-4cce-beca-382f73ae28c0 service nova] Lock "084efb1b-dded-490b-a3fb-150417fe7588-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1369.990109] env[65522]: DEBUG oslo_concurrency.lockutils [req-87598f4a-96ba-4134-bbce-e53d042a294e req-31612da2-c92c-4cce-beca-382f73ae28c0 service nova] Lock "084efb1b-dded-490b-a3fb-150417fe7588-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1369.990756] env[65522]: DEBUG nova.compute.manager [req-87598f4a-96ba-4134-bbce-e53d042a294e req-31612da2-c92c-4cce-beca-382f73ae28c0 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] No waiting events found dispatching network-vif-unplugged-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1369.990756] env[65522]: WARNING nova.compute.manager [req-87598f4a-96ba-4134-bbce-e53d042a294e req-31612da2-c92c-4cce-beca-382f73ae28c0 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Received unexpected event network-vif-unplugged-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed for instance with vm_state shelved and task_state shelving_offloading. [ 1370.084561] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1370.085580] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b791f53-99b7-40a2-baae-24f15b028e67 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.094395] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1370.094651] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03670512-cf3e-4e05-8848-77a7d76d05e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.156171] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1370.156444] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1370.156646] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleting the datastore file [datastore1] 084efb1b-dded-490b-a3fb-150417fe7588 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1370.156933] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c61a3b5-bcfc-4e31-95c4-ace4a52fc78c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.164367] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1370.164367] env[65522]: value = "task-5115245" [ 1370.164367] env[65522]: _type = "Task" [ 1370.164367] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1370.172681] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115245, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1370.674553] env[65522]: DEBUG oslo_vmware.api [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115245, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164314} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1370.674742] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1370.675024] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1370.675141] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1370.727217] env[65522]: INFO nova.scheduler.client.report [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleted allocations for instance 084efb1b-dded-490b-a3fb-150417fe7588 [ 1371.231802] env[65522]: DEBUG oslo_concurrency.lockutils [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1371.232218] env[65522]: DEBUG oslo_concurrency.lockutils [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1371.232340] env[65522]: DEBUG nova.objects.instance [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'resources' on Instance uuid 084efb1b-dded-490b-a3fb-150417fe7588 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1371.734677] env[65522]: DEBUG nova.objects.instance [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'numa_topology' on Instance uuid 084efb1b-dded-490b-a3fb-150417fe7588 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1372.025378] env[65522]: DEBUG nova.compute.manager [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Received event network-changed-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1372.025566] env[65522]: DEBUG nova.compute.manager [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Refreshing instance network info cache due to event network-changed-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1372.025781] env[65522]: DEBUG oslo_concurrency.lockutils [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] Acquiring lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1372.025920] env[65522]: DEBUG oslo_concurrency.lockutils [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] Acquired lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1372.026090] env[65522]: DEBUG nova.network.neutron [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Refreshing network info cache for port 6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1372.237642] env[65522]: DEBUG nova.objects.base [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Object Instance<084efb1b-dded-490b-a3fb-150417fe7588> lazy-loaded attributes: resources,numa_topology {{(pid=65522) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1372.263228] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff466d0-2258-41f6-813c-4bd6247c1674 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.271518] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-941c0d55-3842-4559-a967-82b70d791972 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.304233] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374f978b-6346-4967-8fe0-d9fdf14e3967 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.311967] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c921e05-38c4-441e-a182-27835cb481aa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.101699] env[65522]: WARNING neutronclient.v2_0.client [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1373.102351] env[65522]: WARNING openstack [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1373.102675] env[65522]: WARNING openstack [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1373.123503] env[65522]: DEBUG nova.compute.provider_tree [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1373.214564] env[65522]: WARNING openstack [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1373.214949] env[65522]: WARNING openstack [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1373.274517] env[65522]: WARNING neutronclient.v2_0.client [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1373.275191] env[65522]: WARNING openstack [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1373.275557] env[65522]: WARNING openstack [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1373.355578] env[65522]: DEBUG nova.network.neutron [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Updated VIF entry in instance network info cache for port 6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1373.355947] env[65522]: DEBUG nova.network.neutron [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Updating instance_info_cache with network_info: [{"id": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "address": "fa:16:3e:87:bc:d7", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap6071459f-85", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1373.614679] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "084efb1b-dded-490b-a3fb-150417fe7588" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1373.626935] env[65522]: DEBUG nova.scheduler.client.report [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1373.858486] env[65522]: DEBUG oslo_concurrency.lockutils [req-d25e10ed-ab6c-45c5-a422-ae0f5ac7ab3b req-c5bb8bb7-ea93-4a29-988c-292098b76421 service nova] Releasing lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1374.131517] env[65522]: DEBUG oslo_concurrency.lockutils [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.899s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1374.639972] env[65522]: DEBUG oslo_concurrency.lockutils [None req-67c9bc19-3f18-4da5-a605-c7c688b1ded4 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "084efb1b-dded-490b-a3fb-150417fe7588" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 22.924s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1374.640935] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "084efb1b-dded-490b-a3fb-150417fe7588" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.026s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1374.641153] env[65522]: INFO nova.compute.manager [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Unshelving [ 1375.668067] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1375.668369] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1375.668561] env[65522]: DEBUG nova.objects.instance [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'pci_requests' on Instance uuid 084efb1b-dded-490b-a3fb-150417fe7588 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1376.172824] env[65522]: DEBUG nova.objects.instance [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'numa_topology' on Instance uuid 084efb1b-dded-490b-a3fb-150417fe7588 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1376.675546] env[65522]: INFO nova.compute.claims [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1377.711666] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e713be8-3df0-4293-8044-eed260e6585c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.719792] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a5d9f7-ae35-48e3-b1cf-673b0002bade {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.750465] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8569e63-674f-423f-8d4d-d268f062ce8c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.757864] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df74c5cc-ce6d-4666-87a6-bf6feea8d17b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.771076] env[65522]: DEBUG nova.compute.provider_tree [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1378.273878] env[65522]: DEBUG nova.scheduler.client.report [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1378.779215] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.111s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1378.779585] env[65522]: WARNING neutronclient.v2_0.client [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1378.810036] env[65522]: INFO nova.network.neutron [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Updating port 6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1380.244217] env[65522]: DEBUG nova.compute.manager [req-e78aca98-20e2-482d-8c79-d34a1694dd6d req-96aa2f2c-6839-409a-a11e-ca2fea73a3d2 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Received event network-vif-plugged-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1380.244507] env[65522]: DEBUG oslo_concurrency.lockutils [req-e78aca98-20e2-482d-8c79-d34a1694dd6d req-96aa2f2c-6839-409a-a11e-ca2fea73a3d2 service nova] Acquiring lock "084efb1b-dded-490b-a3fb-150417fe7588-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1380.244670] env[65522]: DEBUG oslo_concurrency.lockutils [req-e78aca98-20e2-482d-8c79-d34a1694dd6d req-96aa2f2c-6839-409a-a11e-ca2fea73a3d2 service nova] Lock "084efb1b-dded-490b-a3fb-150417fe7588-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1380.244846] env[65522]: DEBUG oslo_concurrency.lockutils [req-e78aca98-20e2-482d-8c79-d34a1694dd6d req-96aa2f2c-6839-409a-a11e-ca2fea73a3d2 service nova] Lock "084efb1b-dded-490b-a3fb-150417fe7588-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1380.245435] env[65522]: DEBUG nova.compute.manager [req-e78aca98-20e2-482d-8c79-d34a1694dd6d req-96aa2f2c-6839-409a-a11e-ca2fea73a3d2 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] No waiting events found dispatching network-vif-plugged-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1380.245609] env[65522]: WARNING nova.compute.manager [req-e78aca98-20e2-482d-8c79-d34a1694dd6d req-96aa2f2c-6839-409a-a11e-ca2fea73a3d2 service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Received unexpected event network-vif-plugged-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed for instance with vm_state shelved_offloaded and task_state spawning. [ 1380.328675] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1380.328841] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1380.329300] env[65522]: DEBUG nova.network.neutron [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1380.832503] env[65522]: WARNING neutronclient.v2_0.client [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1380.833030] env[65522]: WARNING openstack [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1380.833470] env[65522]: WARNING openstack [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1380.950206] env[65522]: WARNING openstack [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1380.950619] env[65522]: WARNING openstack [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1381.015498] env[65522]: WARNING neutronclient.v2_0.client [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1381.016178] env[65522]: WARNING openstack [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1381.016614] env[65522]: WARNING openstack [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1381.101787] env[65522]: DEBUG nova.network.neutron [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Updating instance_info_cache with network_info: [{"id": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "address": "fa:16:3e:87:bc:d7", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6071459f-85", "ovs_interfaceid": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1381.605290] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1381.631975] env[65522]: DEBUG nova.virt.hardware [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='598a0728bd6e912fecfc5b551d08243f',container_format='bare',created_at=2025-12-12T19:10:55Z,direct_url=,disk_format='vmdk',id=bbfe6c9f-9556-4717-8ecd-51b706a27303,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-2018576367-shelved',owner='0d1239b79ae94cceb89ae7a8bd57da08',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2025-12-12T19:11:10Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1381.632259] env[65522]: DEBUG nova.virt.hardware [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1381.632412] env[65522]: DEBUG nova.virt.hardware [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1381.632591] env[65522]: DEBUG nova.virt.hardware [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1381.632728] env[65522]: DEBUG nova.virt.hardware [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1381.632867] env[65522]: DEBUG nova.virt.hardware [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1381.633081] env[65522]: DEBUG nova.virt.hardware [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1381.633269] env[65522]: DEBUG nova.virt.hardware [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1381.633463] env[65522]: DEBUG nova.virt.hardware [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1381.633626] env[65522]: DEBUG nova.virt.hardware [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1381.633791] env[65522]: DEBUG nova.virt.hardware [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1381.634952] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86473245-0d71-4947-a110-710b40ad2786 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.643641] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-241e4d71-007f-454f-8552-84bdc294c376 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.656910] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:bc:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a91c3a96-63d0-407c-bcde-c3d5b58d9cb2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1381.664012] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1381.664259] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1381.664466] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-167a5a76-fb94-43ec-9f3a-c536ba8232f7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.684009] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1381.684009] env[65522]: value = "task-5115246" [ 1381.684009] env[65522]: _type = "Task" [ 1381.684009] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1381.691889] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115246, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1382.194819] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115246, 'name': CreateVM_Task, 'duration_secs': 0.340623} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1382.195054] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1382.195528] env[65522]: WARNING neutronclient.v2_0.client [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1382.196047] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bbfe6c9f-9556-4717-8ecd-51b706a27303" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1382.196047] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bbfe6c9f-9556-4717-8ecd-51b706a27303" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1382.196418] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/bbfe6c9f-9556-4717-8ecd-51b706a27303" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1382.196680] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b50e8452-aefe-4ab2-80ba-fed6e70d43b7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.202258] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1382.202258] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52480311-c610-6b4c-9dbf-47d9c9b89c61" [ 1382.202258] env[65522]: _type = "Task" [ 1382.202258] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1382.212341] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52480311-c610-6b4c-9dbf-47d9c9b89c61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1382.274400] env[65522]: DEBUG nova.compute.manager [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Received event network-changed-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1382.274400] env[65522]: DEBUG nova.compute.manager [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Refreshing instance network info cache due to event network-changed-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1382.274400] env[65522]: DEBUG oslo_concurrency.lockutils [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] Acquiring lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1382.274400] env[65522]: DEBUG oslo_concurrency.lockutils [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] Acquired lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1382.274760] env[65522]: DEBUG nova.network.neutron [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Refreshing network info cache for port 6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1382.713145] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bbfe6c9f-9556-4717-8ecd-51b706a27303" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1382.713580] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Processing image bbfe6c9f-9556-4717-8ecd-51b706a27303 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1382.713623] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/bbfe6c9f-9556-4717-8ecd-51b706a27303/bbfe6c9f-9556-4717-8ecd-51b706a27303.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1382.713748] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquired lock "[datastore1] devstack-image-cache_base/bbfe6c9f-9556-4717-8ecd-51b706a27303/bbfe6c9f-9556-4717-8ecd-51b706a27303.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1382.714079] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1382.714270] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-edbd4969-6de6-4a6a-bf9a-5c42a2b0f8ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.723627] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1382.723830] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1382.724590] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af2ad0c7-0dba-4802-9158-079e522f8972 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1382.730217] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1382.730217] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52adcca6-c08e-9653-e007-9229d3e9d3e0" [ 1382.730217] env[65522]: _type = "Task" [ 1382.730217] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1382.739959] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52adcca6-c08e-9653-e007-9229d3e9d3e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1382.776820] env[65522]: WARNING neutronclient.v2_0.client [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1382.777519] env[65522]: WARNING openstack [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1382.777914] env[65522]: WARNING openstack [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1382.885505] env[65522]: WARNING openstack [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1382.885891] env[65522]: WARNING openstack [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1382.945531] env[65522]: WARNING neutronclient.v2_0.client [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1382.947041] env[65522]: WARNING openstack [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1382.947041] env[65522]: WARNING openstack [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1383.029060] env[65522]: DEBUG nova.network.neutron [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Updated VIF entry in instance network info cache for port 6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1383.029453] env[65522]: DEBUG nova.network.neutron [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Updating instance_info_cache with network_info: [{"id": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "address": "fa:16:3e:87:bc:d7", "network": {"id": "d52f9ec8-d771-4bcf-a63a-951baf871ca2", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1287194181-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d1239b79ae94cceb89ae7a8bd57da08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a91c3a96-63d0-407c-bcde-c3d5b58d9cb2", "external-id": "nsx-vlan-transportzone-170", "segmentation_id": 170, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6071459f-85", "ovs_interfaceid": "6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1383.240582] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Preparing fetch location {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1383.240858] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Fetch image to [datastore1] OSTACK_IMG_03e50cc9-aca5-4a53-8d7e-8cf3e8b6f013/OSTACK_IMG_03e50cc9-aca5-4a53-8d7e-8cf3e8b6f013.vmdk {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1383.241129] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Downloading stream optimized image bbfe6c9f-9556-4717-8ecd-51b706a27303 to [datastore1] OSTACK_IMG_03e50cc9-aca5-4a53-8d7e-8cf3e8b6f013/OSTACK_IMG_03e50cc9-aca5-4a53-8d7e-8cf3e8b6f013.vmdk on the data store datastore1 as vApp {{(pid=65522) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1383.241346] env[65522]: DEBUG nova.virt.vmwareapi.images [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Downloading image file data bbfe6c9f-9556-4717-8ecd-51b706a27303 to the ESX as VM named 'OSTACK_IMG_03e50cc9-aca5-4a53-8d7e-8cf3e8b6f013' {{(pid=65522) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1383.311472] env[65522]: DEBUG oslo_vmware.rw_handles [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1383.311472] env[65522]: value = "resgroup-9" [ 1383.311472] env[65522]: _type = "ResourcePool" [ 1383.311472] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1383.311800] env[65522]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-f6cd19b1-b15b-49e6-a11b-11a4a3989912 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.334103] env[65522]: DEBUG oslo_vmware.rw_handles [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lease: (returnval){ [ 1383.334103] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b70837-7984-4a1c-b147-c34188b5fadf" [ 1383.334103] env[65522]: _type = "HttpNfcLease" [ 1383.334103] env[65522]: } obtained for vApp import into resource pool (val){ [ 1383.334103] env[65522]: value = "resgroup-9" [ 1383.334103] env[65522]: _type = "ResourcePool" [ 1383.334103] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1383.334504] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the lease: (returnval){ [ 1383.334504] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b70837-7984-4a1c-b147-c34188b5fadf" [ 1383.334504] env[65522]: _type = "HttpNfcLease" [ 1383.334504] env[65522]: } to be ready. {{(pid=65522) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1383.343444] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1383.343444] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b70837-7984-4a1c-b147-c34188b5fadf" [ 1383.343444] env[65522]: _type = "HttpNfcLease" [ 1383.343444] env[65522]: } is initializing. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1383.532263] env[65522]: DEBUG oslo_concurrency.lockutils [req-ad05d143-95e3-48b6-be47-4482f1b45f07 req-3b82b18b-820c-4dc5-a5aa-b3c1290a8f7d service nova] Releasing lock "refresh_cache-084efb1b-dded-490b-a3fb-150417fe7588" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1383.843019] env[65522]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1383.843019] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b70837-7984-4a1c-b147-c34188b5fadf" [ 1383.843019] env[65522]: _type = "HttpNfcLease" [ 1383.843019] env[65522]: } is ready. {{(pid=65522) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1383.843538] env[65522]: DEBUG oslo_vmware.rw_handles [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1383.843538] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52b70837-7984-4a1c-b147-c34188b5fadf" [ 1383.843538] env[65522]: _type = "HttpNfcLease" [ 1383.843538] env[65522]: }. {{(pid=65522) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1383.844118] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6cec90c-c00c-46ca-ab39-17b505fc87f0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1383.852052] env[65522]: DEBUG oslo_vmware.rw_handles [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e20a15-dc09-c60a-3957-536b06819bc5/disk-0.vmdk from lease info. {{(pid=65522) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1383.852257] env[65522]: DEBUG oslo_vmware.rw_handles [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e20a15-dc09-c60a-3957-536b06819bc5/disk-0.vmdk. {{(pid=65522) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1383.917174] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b9d03ea0-5fb6-49ac-a749-8a0593d47da7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.054328] env[65522]: DEBUG oslo_vmware.rw_handles [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Completed reading data from the image iterator. {{(pid=65522) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1385.054755] env[65522]: DEBUG oslo_vmware.rw_handles [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e20a15-dc09-c60a-3957-536b06819bc5/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1385.055750] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e94f6c-f51b-4dc2-8acb-5d5bbe2318da {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.063849] env[65522]: DEBUG oslo_vmware.rw_handles [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e20a15-dc09-c60a-3957-536b06819bc5/disk-0.vmdk is in state: ready. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1385.064081] env[65522]: DEBUG oslo_vmware.rw_handles [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e20a15-dc09-c60a-3957-536b06819bc5/disk-0.vmdk. {{(pid=65522) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1385.064341] env[65522]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-55286fa8-5aa1-480e-9804-3d886e3f9749 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.237081] env[65522]: DEBUG oslo_vmware.rw_handles [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e20a15-dc09-c60a-3957-536b06819bc5/disk-0.vmdk. {{(pid=65522) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1385.237314] env[65522]: INFO nova.virt.vmwareapi.images [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Downloaded image file data bbfe6c9f-9556-4717-8ecd-51b706a27303 [ 1385.238219] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80477f9b-c492-459d-b729-30272a524546 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.254555] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-04f2ed12-82e8-49f5-8299-5d0f3a2e5dfa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.283734] env[65522]: INFO nova.virt.vmwareapi.images [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] The imported VM was unregistered [ 1385.286745] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Caching image {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1385.286992] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Creating directory with path [datastore1] devstack-image-cache_base/bbfe6c9f-9556-4717-8ecd-51b706a27303 {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1385.287308] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fea1edc2-0459-410a-be90-be59d7fe3241 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.312193] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Created directory with path [datastore1] devstack-image-cache_base/bbfe6c9f-9556-4717-8ecd-51b706a27303 {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1385.312399] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_03e50cc9-aca5-4a53-8d7e-8cf3e8b6f013/OSTACK_IMG_03e50cc9-aca5-4a53-8d7e-8cf3e8b6f013.vmdk to [datastore1] devstack-image-cache_base/bbfe6c9f-9556-4717-8ecd-51b706a27303/bbfe6c9f-9556-4717-8ecd-51b706a27303.vmdk. {{(pid=65522) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1385.312683] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-52e050b1-f7d1-4e8b-8b2e-597fcb9080a7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.320508] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1385.320508] env[65522]: value = "task-5115249" [ 1385.320508] env[65522]: _type = "Task" [ 1385.320508] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1385.329410] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115249, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1385.832195] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115249, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1386.333081] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115249, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1386.835239] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115249, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1387.334218] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115249, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1387.833426] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115249, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1388.334738] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115249, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.523272} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1388.335158] env[65522]: INFO nova.virt.vmwareapi.ds_util [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_03e50cc9-aca5-4a53-8d7e-8cf3e8b6f013/OSTACK_IMG_03e50cc9-aca5-4a53-8d7e-8cf3e8b6f013.vmdk to [datastore1] devstack-image-cache_base/bbfe6c9f-9556-4717-8ecd-51b706a27303/bbfe6c9f-9556-4717-8ecd-51b706a27303.vmdk. [ 1388.335201] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Cleaning up location [datastore1] OSTACK_IMG_03e50cc9-aca5-4a53-8d7e-8cf3e8b6f013 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1388.335383] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_03e50cc9-aca5-4a53-8d7e-8cf3e8b6f013 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1388.335653] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e589d819-c091-428d-a310-be94b2f26a2c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.342758] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1388.342758] env[65522]: value = "task-5115250" [ 1388.342758] env[65522]: _type = "Task" [ 1388.342758] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1388.353850] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115250, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1388.853304] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115250, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.044111} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1388.853502] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1388.853659] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Releasing lock "[datastore1] devstack-image-cache_base/bbfe6c9f-9556-4717-8ecd-51b706a27303/bbfe6c9f-9556-4717-8ecd-51b706a27303.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1388.853899] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/bbfe6c9f-9556-4717-8ecd-51b706a27303/bbfe6c9f-9556-4717-8ecd-51b706a27303.vmdk to [datastore1] 084efb1b-dded-490b-a3fb-150417fe7588/084efb1b-dded-490b-a3fb-150417fe7588.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1388.854292] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eea26b13-9a6a-4ebe-a275-851e68e9fb2f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1388.862173] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1388.862173] env[65522]: value = "task-5115251" [ 1388.862173] env[65522]: _type = "Task" [ 1388.862173] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1388.871161] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115251, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1389.374455] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115251, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1389.876372] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115251, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1390.376040] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115251, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1390.878490] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115251, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1391.377303] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115251, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.384293} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1391.377778] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/bbfe6c9f-9556-4717-8ecd-51b706a27303/bbfe6c9f-9556-4717-8ecd-51b706a27303.vmdk to [datastore1] 084efb1b-dded-490b-a3fb-150417fe7588/084efb1b-dded-490b-a3fb-150417fe7588.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1391.378431] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0f7d96-9628-4ecd-926c-0ecfa94b3c4b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.401534] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Reconfiguring VM instance instance-00000079 to attach disk [datastore1] 084efb1b-dded-490b-a3fb-150417fe7588/084efb1b-dded-490b-a3fb-150417fe7588.vmdk or device None with type streamOptimized {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1391.401890] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-765cd0a5-b843-4377-a0d5-21fa690ebcf1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.423146] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1391.423146] env[65522]: value = "task-5115252" [ 1391.423146] env[65522]: _type = "Task" [ 1391.423146] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1391.431625] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115252, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1391.934074] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115252, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1392.434841] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115252, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1392.935025] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115252, 'name': ReconfigVM_Task, 'duration_secs': 1.300123} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1392.935320] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Reconfigured VM instance instance-00000079 to attach disk [datastore1] 084efb1b-dded-490b-a3fb-150417fe7588/084efb1b-dded-490b-a3fb-150417fe7588.vmdk or device None with type streamOptimized {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1392.935989] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2ae8ada5-5832-4e86-b8c7-623802d18024 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.943366] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1392.943366] env[65522]: value = "task-5115253" [ 1392.943366] env[65522]: _type = "Task" [ 1392.943366] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1392.951659] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115253, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1393.454474] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115253, 'name': Rename_Task, 'duration_secs': 0.139512} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1393.454801] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1393.455083] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f44fc896-18af-4fc4-931d-51d6b3387447 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1393.461843] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1393.461843] env[65522]: value = "task-5115254" [ 1393.461843] env[65522]: _type = "Task" [ 1393.461843] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1393.469713] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115254, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1393.972518] env[65522]: DEBUG oslo_vmware.api [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115254, 'name': PowerOnVM_Task, 'duration_secs': 0.464957} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1393.972864] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1394.073305] env[65522]: DEBUG nova.compute.manager [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1394.074268] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b76077a-5953-43bd-8c36-62e57954c850 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.592465] env[65522]: DEBUG oslo_concurrency.lockutils [None req-53f6ce27-8db7-4240-9c32-0bace2a4c8dd tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "084efb1b-dded-490b-a3fb-150417fe7588" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.951s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1395.977545] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "084efb1b-dded-490b-a3fb-150417fe7588" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1395.977923] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "084efb1b-dded-490b-a3fb-150417fe7588" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1395.978580] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "084efb1b-dded-490b-a3fb-150417fe7588-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1395.978810] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "084efb1b-dded-490b-a3fb-150417fe7588-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1395.979033] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "084efb1b-dded-490b-a3fb-150417fe7588-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1395.981219] env[65522]: INFO nova.compute.manager [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Terminating instance [ 1396.488084] env[65522]: DEBUG nova.compute.manager [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1396.488084] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1396.488654] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13e9993-fe75-4444-952b-10839dcad71f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.497626] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1396.497943] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35c4c6e6-0b76-40d3-888b-0bf1a69421c9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.505978] env[65522]: DEBUG oslo_vmware.api [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1396.505978] env[65522]: value = "task-5115255" [ 1396.505978] env[65522]: _type = "Task" [ 1396.505978] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1396.515344] env[65522]: DEBUG oslo_vmware.api [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115255, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1397.016314] env[65522]: DEBUG oslo_vmware.api [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115255, 'name': PowerOffVM_Task, 'duration_secs': 0.198548} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1397.016733] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1397.016797] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1397.017026] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3db8f043-120d-40bd-b2d2-23f8e7ad2b49 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.077542] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1397.077870] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1397.078185] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleting the datastore file [datastore1] 084efb1b-dded-490b-a3fb-150417fe7588 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1397.078583] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-794a935f-d08a-4569-8506-2351809f543c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.086082] env[65522]: DEBUG oslo_vmware.api [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for the task: (returnval){ [ 1397.086082] env[65522]: value = "task-5115257" [ 1397.086082] env[65522]: _type = "Task" [ 1397.086082] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1397.094794] env[65522]: DEBUG oslo_vmware.api [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115257, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1397.596608] env[65522]: DEBUG oslo_vmware.api [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Task: {'id': task-5115257, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130193} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1397.596955] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1397.597237] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1397.597460] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1397.597666] env[65522]: INFO nova.compute.manager [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1397.597940] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1397.598199] env[65522]: DEBUG nova.compute.manager [-] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1397.598304] env[65522]: DEBUG nova.network.neutron [-] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1397.598588] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1397.599200] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1397.599530] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1397.667332] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1398.130167] env[65522]: DEBUG nova.compute.manager [req-23c3a38c-6be0-4dd2-bdbd-e0082e273576 req-d1a6c3dd-91de-4b2a-b621-37c5a9fc269e service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Received event network-vif-deleted-6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1398.130503] env[65522]: INFO nova.compute.manager [req-23c3a38c-6be0-4dd2-bdbd-e0082e273576 req-d1a6c3dd-91de-4b2a-b621-37c5a9fc269e service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Neutron deleted interface 6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed; detaching it from the instance and deleting it from the info cache [ 1398.130503] env[65522]: DEBUG nova.network.neutron [req-23c3a38c-6be0-4dd2-bdbd-e0082e273576 req-d1a6c3dd-91de-4b2a-b621-37c5a9fc269e service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1398.609065] env[65522]: DEBUG nova.network.neutron [-] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1398.632415] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f0b78269-75d7-4704-93bc-c7db36168035 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.644593] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad291f6d-f2e6-40c9-be3a-adc9e3dd9e8a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1398.670652] env[65522]: DEBUG nova.compute.manager [req-23c3a38c-6be0-4dd2-bdbd-e0082e273576 req-d1a6c3dd-91de-4b2a-b621-37c5a9fc269e service nova] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Detach interface failed, port_id=6071459f-85f6-4d6b-9eb7-3c82d9f1e6ed, reason: Instance 084efb1b-dded-490b-a3fb-150417fe7588 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1399.112112] env[65522]: INFO nova.compute.manager [-] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Took 1.51 seconds to deallocate network for instance. [ 1399.618565] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1399.618965] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1399.619091] env[65522]: DEBUG nova.objects.instance [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lazy-loading 'resources' on Instance uuid 084efb1b-dded-490b-a3fb-150417fe7588 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1400.154177] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c41c9c9-d9f9-4cbc-8c92-4c94b0b3a89e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.161812] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3fbfa54-acc6-4052-9d0c-5e78f6d2e720 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.191081] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b8c642-a418-4748-8628-5e6792470256 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.198939] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63a907c-74bb-453b-80c7-15464e1c1139 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.213480] env[65522]: DEBUG nova.compute.provider_tree [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1400.717027] env[65522]: DEBUG nova.scheduler.client.report [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1401.222040] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.603s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1401.243730] env[65522]: INFO nova.scheduler.client.report [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Deleted allocations for instance 084efb1b-dded-490b-a3fb-150417fe7588 [ 1401.751889] env[65522]: DEBUG oslo_concurrency.lockutils [None req-58d78dd3-f71d-4b80-89ac-0e5fb6c50919 tempest-ServerActionsTestOtherB-1255887243 tempest-ServerActionsTestOtherB-1255887243-project-member] Lock "084efb1b-dded-490b-a3fb-150417fe7588" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.774s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1406.894100] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1406.894560] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.401974] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.402197] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.402350] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.402497] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.402636] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.402841] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.403302] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1407.403302] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.908296] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1407.908675] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1407.908726] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1407.908992] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1407.909998] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166b2f60-d944-4cff-a242-29124b0e5433 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.919072] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfcbad13-58e4-4849-b38a-666627cb376a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.934036] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8697c32e-c111-4775-8b85-16038ca9ac37 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.941483] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d97127e-2815-4560-8f3c-277e78d6c54b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.972433] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180514MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1407.972634] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1407.972854] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1408.993644] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1408.995192] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=100GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] stats={'failed_builds': '0'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1409.010917] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd556bc5-f77a-4407-a001-6bbfc190f792 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.019875] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa956c91-d5cf-40ec-8d7a-71800d0e11d2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.050704] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-105552df-06eb-4970-b422-e2824e8e313f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.059041] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13cea54-ed23-4d49-b48b-e6aed7653576 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.072939] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1409.575940] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1410.082696] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1410.084127] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.110s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1411.527211] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "148f338e-d583-46b7-8ae0-855cb280cd47" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1411.527519] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "148f338e-d583-46b7-8ae0-855cb280cd47" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1412.029583] env[65522]: DEBUG nova.compute.manager [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1412.550031] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1412.550407] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1412.551929] env[65522]: INFO nova.compute.claims [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1413.589568] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19c209a-6993-4afd-a9c0-99d0a609ce31 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.599827] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a8b419-e507-48cc-b656-50d6a3445cf3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.631296] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0edaa98c-9835-4722-b4ce-72127d6464e6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.639675] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e26065c-bd8f-45d2-90da-5666556d96ed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.654595] env[65522]: DEBUG nova.compute.provider_tree [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1414.158346] env[65522]: DEBUG nova.scheduler.client.report [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1414.663107] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.113s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1414.663720] env[65522]: DEBUG nova.compute.manager [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1415.169819] env[65522]: DEBUG nova.compute.utils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1415.171313] env[65522]: DEBUG nova.compute.manager [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1415.171523] env[65522]: DEBUG nova.network.neutron [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1415.171820] env[65522]: WARNING neutronclient.v2_0.client [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1415.172139] env[65522]: WARNING neutronclient.v2_0.client [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1415.172727] env[65522]: WARNING openstack [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1415.173073] env[65522]: WARNING openstack [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1415.232258] env[65522]: DEBUG nova.policy [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f35b3e2725f49e6bd84e0aa787c0e43', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37a72785dad54301a64694398f167f92', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1415.530333] env[65522]: DEBUG nova.network.neutron [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Successfully created port: df1f3178-7f92-496b-8bc7-908acf23200f {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1415.682235] env[65522]: DEBUG nova.compute.manager [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1416.692129] env[65522]: DEBUG nova.compute.manager [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1416.719034] env[65522]: DEBUG nova.virt.hardware [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1416.719315] env[65522]: DEBUG nova.virt.hardware [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1416.719470] env[65522]: DEBUG nova.virt.hardware [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1416.719650] env[65522]: DEBUG nova.virt.hardware [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1416.719795] env[65522]: DEBUG nova.virt.hardware [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1416.719939] env[65522]: DEBUG nova.virt.hardware [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1416.720165] env[65522]: DEBUG nova.virt.hardware [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1416.720349] env[65522]: DEBUG nova.virt.hardware [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1416.720619] env[65522]: DEBUG nova.virt.hardware [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1416.720917] env[65522]: DEBUG nova.virt.hardware [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1416.721206] env[65522]: DEBUG nova.virt.hardware [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1416.722202] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4764154-02d2-47f4-a0de-7a0b90d90c6d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.731212] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd1397c-1ba5-426b-a7b9-b6a88f99af1f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.916861] env[65522]: DEBUG nova.compute.manager [req-e93fa825-6248-4e65-a15a-a29606190d2c req-11a5fe3b-2eac-4dfa-af1c-1a9fb64e6c87 service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Received event network-vif-plugged-df1f3178-7f92-496b-8bc7-908acf23200f {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1416.916861] env[65522]: DEBUG oslo_concurrency.lockutils [req-e93fa825-6248-4e65-a15a-a29606190d2c req-11a5fe3b-2eac-4dfa-af1c-1a9fb64e6c87 service nova] Acquiring lock "148f338e-d583-46b7-8ae0-855cb280cd47-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1416.917073] env[65522]: DEBUG oslo_concurrency.lockutils [req-e93fa825-6248-4e65-a15a-a29606190d2c req-11a5fe3b-2eac-4dfa-af1c-1a9fb64e6c87 service nova] Lock "148f338e-d583-46b7-8ae0-855cb280cd47-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1416.917307] env[65522]: DEBUG oslo_concurrency.lockutils [req-e93fa825-6248-4e65-a15a-a29606190d2c req-11a5fe3b-2eac-4dfa-af1c-1a9fb64e6c87 service nova] Lock "148f338e-d583-46b7-8ae0-855cb280cd47-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1416.917416] env[65522]: DEBUG nova.compute.manager [req-e93fa825-6248-4e65-a15a-a29606190d2c req-11a5fe3b-2eac-4dfa-af1c-1a9fb64e6c87 service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] No waiting events found dispatching network-vif-plugged-df1f3178-7f92-496b-8bc7-908acf23200f {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1416.917579] env[65522]: WARNING nova.compute.manager [req-e93fa825-6248-4e65-a15a-a29606190d2c req-11a5fe3b-2eac-4dfa-af1c-1a9fb64e6c87 service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Received unexpected event network-vif-plugged-df1f3178-7f92-496b-8bc7-908acf23200f for instance with vm_state building and task_state spawning. [ 1417.008677] env[65522]: DEBUG nova.network.neutron [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Successfully updated port: df1f3178-7f92-496b-8bc7-908acf23200f {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1417.512220] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "refresh_cache-148f338e-d583-46b7-8ae0-855cb280cd47" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1417.512356] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired lock "refresh_cache-148f338e-d583-46b7-8ae0-855cb280cd47" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1417.512480] env[65522]: DEBUG nova.network.neutron [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1418.017107] env[65522]: WARNING openstack [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1418.017502] env[65522]: WARNING openstack [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1418.053517] env[65522]: DEBUG nova.network.neutron [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1418.073363] env[65522]: WARNING openstack [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1418.073748] env[65522]: WARNING openstack [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1418.135215] env[65522]: WARNING neutronclient.v2_0.client [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1418.135907] env[65522]: WARNING openstack [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1418.136279] env[65522]: WARNING openstack [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1418.219849] env[65522]: DEBUG nova.network.neutron [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Updating instance_info_cache with network_info: [{"id": "df1f3178-7f92-496b-8bc7-908acf23200f", "address": "fa:16:3e:15:8c:3b", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf1f3178-7f", "ovs_interfaceid": "df1f3178-7f92-496b-8bc7-908acf23200f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1418.723077] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Releasing lock "refresh_cache-148f338e-d583-46b7-8ae0-855cb280cd47" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1418.723444] env[65522]: DEBUG nova.compute.manager [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Instance network_info: |[{"id": "df1f3178-7f92-496b-8bc7-908acf23200f", "address": "fa:16:3e:15:8c:3b", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf1f3178-7f", "ovs_interfaceid": "df1f3178-7f92-496b-8bc7-908acf23200f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1418.723933] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:8c:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '07e9bef1-2b0e-4e4d-997f-de71bb0e213a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'df1f3178-7f92-496b-8bc7-908acf23200f', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1418.731369] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Creating folder: Project (37a72785dad54301a64694398f167f92). Parent ref: group-v994660. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1418.731655] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9e22c7f7-1337-43ca-b5bc-f44c29e47275 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.744378] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Created folder: Project (37a72785dad54301a64694398f167f92) in parent group-v994660. [ 1418.744548] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Creating folder: Instances. Parent ref: group-v994998. {{(pid=65522) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1418.744790] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce2cb905-1463-4f79-be2e-bc6acb39ddf8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.755381] env[65522]: INFO nova.virt.vmwareapi.vm_util [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Created folder: Instances in parent group-v994998. [ 1418.755625] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1418.755866] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1418.756101] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0039483f-a34d-43af-ac10-0bdd632b6319 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.776776] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1418.776776] env[65522]: value = "task-5115261" [ 1418.776776] env[65522]: _type = "Task" [ 1418.776776] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1418.785128] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115261, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1418.946598] env[65522]: DEBUG nova.compute.manager [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Received event network-changed-df1f3178-7f92-496b-8bc7-908acf23200f {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1418.946822] env[65522]: DEBUG nova.compute.manager [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Refreshing instance network info cache due to event network-changed-df1f3178-7f92-496b-8bc7-908acf23200f. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1418.947115] env[65522]: DEBUG oslo_concurrency.lockutils [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] Acquiring lock "refresh_cache-148f338e-d583-46b7-8ae0-855cb280cd47" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1418.947280] env[65522]: DEBUG oslo_concurrency.lockutils [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] Acquired lock "refresh_cache-148f338e-d583-46b7-8ae0-855cb280cd47" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1418.947510] env[65522]: DEBUG nova.network.neutron [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Refreshing network info cache for port df1f3178-7f92-496b-8bc7-908acf23200f {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1419.288661] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115261, 'name': CreateVM_Task, 'duration_secs': 0.317097} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1419.289104] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1419.289397] env[65522]: WARNING neutronclient.v2_0.client [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1419.289959] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1419.289959] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1419.290256] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1419.290562] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f47282fe-5962-46ce-8d77-bc8da5bf4901 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.296700] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1419.296700] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52800b4d-8306-a48a-9cd8-3f0747110fd2" [ 1419.296700] env[65522]: _type = "Task" [ 1419.296700] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1419.305566] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52800b4d-8306-a48a-9cd8-3f0747110fd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1419.450468] env[65522]: WARNING neutronclient.v2_0.client [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1419.451169] env[65522]: WARNING openstack [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1419.451537] env[65522]: WARNING openstack [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1419.546369] env[65522]: WARNING openstack [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1419.546753] env[65522]: WARNING openstack [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1419.605128] env[65522]: WARNING neutronclient.v2_0.client [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1419.605824] env[65522]: WARNING openstack [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1419.606187] env[65522]: WARNING openstack [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1419.687305] env[65522]: DEBUG nova.network.neutron [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Updated VIF entry in instance network info cache for port df1f3178-7f92-496b-8bc7-908acf23200f. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1419.687688] env[65522]: DEBUG nova.network.neutron [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Updating instance_info_cache with network_info: [{"id": "df1f3178-7f92-496b-8bc7-908acf23200f", "address": "fa:16:3e:15:8c:3b", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf1f3178-7f", "ovs_interfaceid": "df1f3178-7f92-496b-8bc7-908acf23200f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1419.807090] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52800b4d-8306-a48a-9cd8-3f0747110fd2, 'name': SearchDatastore_Task, 'duration_secs': 0.010812} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1419.807354] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1419.807596] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1419.807833] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1419.808082] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1419.808288] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1419.808560] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7db079bb-53b6-4516-ad3e-91f5aaf2d0ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.818277] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1419.818427] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1419.819173] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-719728a3-e976-4573-b1fd-cdec83688df3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.824954] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1419.824954] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a8821e-3a3f-f241-dc68-fdcab8ac9f5d" [ 1419.824954] env[65522]: _type = "Task" [ 1419.824954] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1419.833681] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a8821e-3a3f-f241-dc68-fdcab8ac9f5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1420.190345] env[65522]: DEBUG oslo_concurrency.lockutils [req-279d62fe-f2dd-49aa-9091-3f6261dbd628 req-9cbe7ed6-afd1-4cb8-b321-51047bc3da8d service nova] Releasing lock "refresh_cache-148f338e-d583-46b7-8ae0-855cb280cd47" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1420.335320] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a8821e-3a3f-f241-dc68-fdcab8ac9f5d, 'name': SearchDatastore_Task, 'duration_secs': 0.009405} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1420.336181] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7c31109-d198-4005-8b0c-1653a2ce9eee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.342357] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1420.342357] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525c0b2d-20e2-bb5a-299b-7c2a20000a58" [ 1420.342357] env[65522]: _type = "Task" [ 1420.342357] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1420.351445] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525c0b2d-20e2-bb5a-299b-7c2a20000a58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1420.853193] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525c0b2d-20e2-bb5a-299b-7c2a20000a58, 'name': SearchDatastore_Task, 'duration_secs': 0.009846} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1420.853483] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1420.853756] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 148f338e-d583-46b7-8ae0-855cb280cd47/148f338e-d583-46b7-8ae0-855cb280cd47.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1420.854028] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1f57eeff-663c-4ff7-a531-cdbd35a3a8cf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1420.861325] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1420.861325] env[65522]: value = "task-5115262" [ 1420.861325] env[65522]: _type = "Task" [ 1420.861325] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1420.869587] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115262, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1421.371978] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115262, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439932} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1421.372382] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 148f338e-d583-46b7-8ae0-855cb280cd47/148f338e-d583-46b7-8ae0-855cb280cd47.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1421.372476] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1421.372733] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-096daafb-033c-41d4-b192-c2e87b34c2ed {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.379210] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1421.379210] env[65522]: value = "task-5115263" [ 1421.379210] env[65522]: _type = "Task" [ 1421.379210] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1421.386706] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115263, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1421.889523] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115263, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067061} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1421.889830] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1421.890673] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac75d80-218e-4a68-8898-a6ec6afd24d7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.913111] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Reconfiguring VM instance instance-0000007c to attach disk [datastore1] 148f338e-d583-46b7-8ae0-855cb280cd47/148f338e-d583-46b7-8ae0-855cb280cd47.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1421.913423] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d2795d2-c6d7-4dfe-b4df-453d7ca63613 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.932842] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1421.932842] env[65522]: value = "task-5115264" [ 1421.932842] env[65522]: _type = "Task" [ 1421.932842] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1421.940884] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115264, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1422.443522] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115264, 'name': ReconfigVM_Task, 'duration_secs': 0.269858} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1422.443943] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Reconfigured VM instance instance-0000007c to attach disk [datastore1] 148f338e-d583-46b7-8ae0-855cb280cd47/148f338e-d583-46b7-8ae0-855cb280cd47.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1422.444465] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-14c87fda-cee2-4445-abb3-a4639261fc60 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.451560] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1422.451560] env[65522]: value = "task-5115265" [ 1422.451560] env[65522]: _type = "Task" [ 1422.451560] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1422.460434] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115265, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1422.962361] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115265, 'name': Rename_Task, 'duration_secs': 0.167897} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1422.962634] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1422.962921] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5970daf4-8be3-4981-b564-1b1e83c233b7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1422.970231] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1422.970231] env[65522]: value = "task-5115266" [ 1422.970231] env[65522]: _type = "Task" [ 1422.970231] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1422.978750] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115266, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1423.480917] env[65522]: DEBUG oslo_vmware.api [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115266, 'name': PowerOnVM_Task, 'duration_secs': 0.433805} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1423.481256] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1423.481452] env[65522]: INFO nova.compute.manager [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Took 6.79 seconds to spawn the instance on the hypervisor. [ 1423.481649] env[65522]: DEBUG nova.compute.manager [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1423.482535] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067d5ddc-9ad3-41bb-888d-477b17f81c0f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.003265] env[65522]: INFO nova.compute.manager [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Took 11.47 seconds to build instance. [ 1424.506755] env[65522]: DEBUG oslo_concurrency.lockutils [None req-71273cba-71f9-40d6-a46d-8f8b18c47ff6 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "148f338e-d583-46b7-8ae0-855cb280cd47" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.978s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1424.557062] env[65522]: DEBUG nova.compute.manager [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Received event network-changed-df1f3178-7f92-496b-8bc7-908acf23200f {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1424.557062] env[65522]: DEBUG nova.compute.manager [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Refreshing instance network info cache due to event network-changed-df1f3178-7f92-496b-8bc7-908acf23200f. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1424.557062] env[65522]: DEBUG oslo_concurrency.lockutils [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] Acquiring lock "refresh_cache-148f338e-d583-46b7-8ae0-855cb280cd47" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1424.557062] env[65522]: DEBUG oslo_concurrency.lockutils [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] Acquired lock "refresh_cache-148f338e-d583-46b7-8ae0-855cb280cd47" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1424.557255] env[65522]: DEBUG nova.network.neutron [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Refreshing network info cache for port df1f3178-7f92-496b-8bc7-908acf23200f {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1425.059265] env[65522]: WARNING neutronclient.v2_0.client [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1425.060448] env[65522]: WARNING openstack [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1425.060713] env[65522]: WARNING openstack [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1425.173284] env[65522]: WARNING openstack [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1425.173724] env[65522]: WARNING openstack [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1425.233731] env[65522]: WARNING neutronclient.v2_0.client [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1425.234408] env[65522]: WARNING openstack [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1425.234741] env[65522]: WARNING openstack [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1425.314834] env[65522]: DEBUG nova.network.neutron [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Updated VIF entry in instance network info cache for port df1f3178-7f92-496b-8bc7-908acf23200f. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1425.315232] env[65522]: DEBUG nova.network.neutron [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Updating instance_info_cache with network_info: [{"id": "df1f3178-7f92-496b-8bc7-908acf23200f", "address": "fa:16:3e:15:8c:3b", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf1f3178-7f", "ovs_interfaceid": "df1f3178-7f92-496b-8bc7-908acf23200f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1425.818337] env[65522]: DEBUG oslo_concurrency.lockutils [req-ca038cb2-1a53-4d81-9565-ca7aac996f85 req-20337a07-ca85-4975-b4fa-58d8f958f6e2 service nova] Releasing lock "refresh_cache-148f338e-d583-46b7-8ae0-855cb280cd47" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1456.150217] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1457.653427] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1457.653824] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1459.145884] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1459.149582] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1460.150054] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1463.150556] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1463.150989] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1463.150989] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1463.151154] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Cleaning up deleted instances with incomplete migration {{(pid=65522) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11971}} [ 1463.738827] env[65522]: DEBUG oslo_concurrency.lockutils [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "148f338e-d583-46b7-8ae0-855cb280cd47" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1463.739073] env[65522]: DEBUG oslo_concurrency.lockutils [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "148f338e-d583-46b7-8ae0-855cb280cd47" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1464.242333] env[65522]: DEBUG nova.compute.utils [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1464.653269] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1464.745302] env[65522]: DEBUG oslo_concurrency.lockutils [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "148f338e-d583-46b7-8ae0-855cb280cd47" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1465.156810] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1465.157122] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1465.157329] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1465.157491] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1465.158430] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09bc985c-1a00-4a35-aa35-8c95472dd06a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.168627] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b212a89-3c60-4040-9ec8-90f5ebf428f4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.183134] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b810cb36-d24f-4ca2-b135-286efa5bbb2f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.190551] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac8d3f8a-2372-498f-94cf-68ac0f125434 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.219787] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180527MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1465.219980] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1465.220265] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1465.808372] env[65522]: DEBUG oslo_concurrency.lockutils [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "148f338e-d583-46b7-8ae0-855cb280cd47" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1465.808775] env[65522]: DEBUG oslo_concurrency.lockutils [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "148f338e-d583-46b7-8ae0-855cb280cd47" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1465.808871] env[65522]: INFO nova.compute.manager [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Attaching volume 384aa72c-340c-4358-8b76-0c62fe61d3aa to /dev/sdb [ 1465.840766] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee02b3c2-6914-48e3-b9b4-7bcb0551adfb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.848291] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf632bc-406a-4109-a4c8-f5038a8f7f4c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.861667] env[65522]: DEBUG nova.virt.block_device [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Updating existing volume attachment record: 1dd3912e-d862-42c5-91b0-93436bff2f0e {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1466.317327] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 148f338e-d583-46b7-8ae0-855cb280cd47 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1466.317583] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1466.317736] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=100GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '1', 'num_vm_active': '1', 'num_task_None': '1', 'num_os_type_None': '1', 'num_proj_37a72785dad54301a64694398f167f92': '1', 'io_workload': '0'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1466.346456] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec8036d-e552-4433-8ebc-307b8fe7a4bf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1466.354496] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d296daa-e986-462b-b82b-52927245acdf {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1466.387715] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a262a63-dd27-4f99-862f-ccb458e36214 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1466.395688] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940308b8-5fc3-464c-84b2-b7836ffd3ac4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1466.410365] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1466.914074] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1467.419030] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1467.419202] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.199s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1468.916611] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1470.907066] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1470.907371] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995002', 'volume_id': '384aa72c-340c-4358-8b76-0c62fe61d3aa', 'name': 'volume-384aa72c-340c-4358-8b76-0c62fe61d3aa', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '148f338e-d583-46b7-8ae0-855cb280cd47', 'attached_at': '', 'detached_at': '', 'volume_id': '384aa72c-340c-4358-8b76-0c62fe61d3aa', 'serial': '384aa72c-340c-4358-8b76-0c62fe61d3aa'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1470.908263] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30e34fb-2c09-4480-8bc2-aa48fc0c1bb7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.925914] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ded08db-6c86-4d1c-894b-9d1104ecb8cc {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.950411] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Reconfiguring VM instance instance-0000007c to attach disk [datastore1] volume-384aa72c-340c-4358-8b76-0c62fe61d3aa/volume-384aa72c-340c-4358-8b76-0c62fe61d3aa.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1470.950680] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4a8551d-e09d-498f-a49e-d0e3e1322aa1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.968600] env[65522]: DEBUG oslo_vmware.api [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1470.968600] env[65522]: value = "task-5115271" [ 1470.968600] env[65522]: _type = "Task" [ 1470.968600] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1470.976717] env[65522]: DEBUG oslo_vmware.api [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115271, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1471.479060] env[65522]: DEBUG oslo_vmware.api [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115271, 'name': ReconfigVM_Task, 'duration_secs': 0.339851} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1471.479363] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Reconfigured VM instance instance-0000007c to attach disk [datastore1] volume-384aa72c-340c-4358-8b76-0c62fe61d3aa/volume-384aa72c-340c-4358-8b76-0c62fe61d3aa.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1471.484349] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa4aa5c2-1d0a-4855-9a88-ae393f4204a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1471.499502] env[65522]: DEBUG oslo_vmware.api [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1471.499502] env[65522]: value = "task-5115272" [ 1471.499502] env[65522]: _type = "Task" [ 1471.499502] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1471.508068] env[65522]: DEBUG oslo_vmware.api [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115272, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1472.010615] env[65522]: DEBUG oslo_vmware.api [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115272, 'name': ReconfigVM_Task, 'duration_secs': 0.143999} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1472.011044] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995002', 'volume_id': '384aa72c-340c-4358-8b76-0c62fe61d3aa', 'name': 'volume-384aa72c-340c-4358-8b76-0c62fe61d3aa', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '148f338e-d583-46b7-8ae0-855cb280cd47', 'attached_at': '', 'detached_at': '', 'volume_id': '384aa72c-340c-4358-8b76-0c62fe61d3aa', 'serial': '384aa72c-340c-4358-8b76-0c62fe61d3aa'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1473.052851] env[65522]: DEBUG nova.objects.instance [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lazy-loading 'flavor' on Instance uuid 148f338e-d583-46b7-8ae0-855cb280cd47 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1473.149700] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1473.149958] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Cleaning up deleted instances {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11933}} [ 1473.557345] env[65522]: DEBUG oslo_concurrency.lockutils [None req-86899bda-6cf3-4300-9d7f-7f7b61ef6bb5 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "148f338e-d583-46b7-8ae0-855cb280cd47" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.749s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1473.657130] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] There are 20 instances to clean {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11942}} [ 1473.657323] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: b56fc96d-bad2-4655-b77f-66674a0c3330] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1474.160994] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: b0b0ca78-65a1-4aca-bb3a-83fe5dd1e187] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1474.400060] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "2ad40e32-4111-4f49-9080-04768d870339" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1474.400060] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "2ad40e32-4111-4f49-9080-04768d870339" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1474.664998] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 084efb1b-dded-490b-a3fb-150417fe7588] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1474.902431] env[65522]: DEBUG nova.compute.manager [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1475.168314] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: bef9977d-ae21-426e-97d3-7b70a7276356] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1475.423343] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1475.423638] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1475.425381] env[65522]: INFO nova.compute.claims [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1475.672032] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: c1aca457-dc7c-477c-9b4f-3ed577fb2adb] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1476.175319] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 874813b4-d0d6-4559-82e2-9fa2ea4feb15] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1476.470781] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d873554-c838-4646-a030-a897485d35b8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.478780] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5312e24b-4cd7-4c76-b9ce-dabb3d82ad4c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.510573] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-180a59ee-18e2-4f7d-bc2d-c377e506b195 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.518525] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad924a4-6aa2-485e-87f3-44002c15d6fa {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1476.531701] env[65522]: DEBUG nova.compute.provider_tree [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1476.678598] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 1b0fcdbb-aa9e-4915-a8c6-1167aa888390] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1477.035100] env[65522]: DEBUG nova.scheduler.client.report [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1477.181666] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: daffdd40-9ee3-4a80-b670-5772395a32ae] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1477.540077] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.116s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1477.540629] env[65522]: DEBUG nova.compute.manager [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1477.685216] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 3adfefa5-fb19-44b2-b3c5-42f2e2918673] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1478.045618] env[65522]: DEBUG nova.compute.utils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1478.047157] env[65522]: DEBUG nova.compute.manager [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1478.047446] env[65522]: DEBUG nova.network.neutron [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1478.047786] env[65522]: WARNING neutronclient.v2_0.client [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1478.048103] env[65522]: WARNING neutronclient.v2_0.client [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1478.048936] env[65522]: WARNING openstack [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1478.049150] env[65522]: WARNING openstack [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1478.105197] env[65522]: DEBUG nova.policy [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f35b3e2725f49e6bd84e0aa787c0e43', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37a72785dad54301a64694398f167f92', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1478.188173] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 874b6a07-ebba-4f75-937e-2281af1598b8] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1478.435086] env[65522]: DEBUG nova.network.neutron [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Successfully created port: e817d4b6-77e2-43b4-932b-9ba3ecd832ef {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1478.557748] env[65522]: DEBUG nova.compute.manager [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1478.691978] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: ae3455d8-2f38-42ad-b16d-d98aef92b2a8] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1479.195591] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 3987d71f-47ab-4dd4-8426-95d78503effa] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1479.569348] env[65522]: DEBUG nova.compute.manager [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1479.598941] env[65522]: DEBUG nova.virt.hardware [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1479.599214] env[65522]: DEBUG nova.virt.hardware [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1479.599371] env[65522]: DEBUG nova.virt.hardware [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1479.599559] env[65522]: DEBUG nova.virt.hardware [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1479.599701] env[65522]: DEBUG nova.virt.hardware [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1479.599847] env[65522]: DEBUG nova.virt.hardware [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1479.600070] env[65522]: DEBUG nova.virt.hardware [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1479.600235] env[65522]: DEBUG nova.virt.hardware [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1479.600402] env[65522]: DEBUG nova.virt.hardware [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1479.600563] env[65522]: DEBUG nova.virt.hardware [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1479.600737] env[65522]: DEBUG nova.virt.hardware [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1479.601634] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502e1a0a-748a-4383-944a-2087e48f2bb4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1479.610227] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b296295-251f-4abe-95a2-985de3fdf480 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1479.698791] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: aebb7082-f88c-42f3-9d9f-93cb2a6f545b] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1479.836728] env[65522]: DEBUG nova.compute.manager [req-0713d1b3-ebf5-4399-9d0c-6f373d71a6a0 req-dd530c29-526d-4284-8352-5fb7e58c9c91 service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Received event network-vif-plugged-e817d4b6-77e2-43b4-932b-9ba3ecd832ef {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1479.836982] env[65522]: DEBUG oslo_concurrency.lockutils [req-0713d1b3-ebf5-4399-9d0c-6f373d71a6a0 req-dd530c29-526d-4284-8352-5fb7e58c9c91 service nova] Acquiring lock "2ad40e32-4111-4f49-9080-04768d870339-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1479.837324] env[65522]: DEBUG oslo_concurrency.lockutils [req-0713d1b3-ebf5-4399-9d0c-6f373d71a6a0 req-dd530c29-526d-4284-8352-5fb7e58c9c91 service nova] Lock "2ad40e32-4111-4f49-9080-04768d870339-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1479.837442] env[65522]: DEBUG oslo_concurrency.lockutils [req-0713d1b3-ebf5-4399-9d0c-6f373d71a6a0 req-dd530c29-526d-4284-8352-5fb7e58c9c91 service nova] Lock "2ad40e32-4111-4f49-9080-04768d870339-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1479.837620] env[65522]: DEBUG nova.compute.manager [req-0713d1b3-ebf5-4399-9d0c-6f373d71a6a0 req-dd530c29-526d-4284-8352-5fb7e58c9c91 service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] No waiting events found dispatching network-vif-plugged-e817d4b6-77e2-43b4-932b-9ba3ecd832ef {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1479.837807] env[65522]: WARNING nova.compute.manager [req-0713d1b3-ebf5-4399-9d0c-6f373d71a6a0 req-dd530c29-526d-4284-8352-5fb7e58c9c91 service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Received unexpected event network-vif-plugged-e817d4b6-77e2-43b4-932b-9ba3ecd832ef for instance with vm_state building and task_state spawning. [ 1479.919941] env[65522]: DEBUG nova.network.neutron [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Successfully updated port: e817d4b6-77e2-43b4-932b-9ba3ecd832ef {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1480.202131] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 6d856111-81bd-4e2f-b7fc-e169c3e974a6] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1480.423327] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "refresh_cache-2ad40e32-4111-4f49-9080-04768d870339" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1480.423450] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired lock "refresh_cache-2ad40e32-4111-4f49-9080-04768d870339" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1480.423627] env[65522]: DEBUG nova.network.neutron [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1480.705024] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 3b82f9ce-ea59-430d-adb6-918cc6cc48a4] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1480.926716] env[65522]: WARNING openstack [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1480.927178] env[65522]: WARNING openstack [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1480.972789] env[65522]: DEBUG nova.network.neutron [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1480.991921] env[65522]: WARNING openstack [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1480.992338] env[65522]: WARNING openstack [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1481.057078] env[65522]: WARNING neutronclient.v2_0.client [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1481.057772] env[65522]: WARNING openstack [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1481.058141] env[65522]: WARNING openstack [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1481.140138] env[65522]: DEBUG nova.network.neutron [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Updating instance_info_cache with network_info: [{"id": "e817d4b6-77e2-43b4-932b-9ba3ecd832ef", "address": "fa:16:3e:1b:28:32", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape817d4b6-77", "ovs_interfaceid": "e817d4b6-77e2-43b4-932b-9ba3ecd832ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1481.208789] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 2981b635-43c8-4bd6-9991-e6af0be82f3c] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1481.643062] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Releasing lock "refresh_cache-2ad40e32-4111-4f49-9080-04768d870339" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1481.643402] env[65522]: DEBUG nova.compute.manager [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Instance network_info: |[{"id": "e817d4b6-77e2-43b4-932b-9ba3ecd832ef", "address": "fa:16:3e:1b:28:32", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape817d4b6-77", "ovs_interfaceid": "e817d4b6-77e2-43b4-932b-9ba3ecd832ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1481.643881] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:28:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '07e9bef1-2b0e-4e4d-997f-de71bb0e213a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e817d4b6-77e2-43b4-932b-9ba3ecd832ef', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1481.651751] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1481.651947] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1481.652570] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d4428445-ddc0-496a-955f-98c21b7ad633 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1481.673548] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1481.673548] env[65522]: value = "task-5115273" [ 1481.673548] env[65522]: _type = "Task" [ 1481.673548] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1481.681677] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115273, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1481.712422] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: cb62df50-2f86-469a-b2ee-fd3754d61c83] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1481.869311] env[65522]: DEBUG nova.compute.manager [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Received event network-changed-e817d4b6-77e2-43b4-932b-9ba3ecd832ef {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1481.869597] env[65522]: DEBUG nova.compute.manager [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Refreshing instance network info cache due to event network-changed-e817d4b6-77e2-43b4-932b-9ba3ecd832ef. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1481.869827] env[65522]: DEBUG oslo_concurrency.lockutils [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] Acquiring lock "refresh_cache-2ad40e32-4111-4f49-9080-04768d870339" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1481.869995] env[65522]: DEBUG oslo_concurrency.lockutils [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] Acquired lock "refresh_cache-2ad40e32-4111-4f49-9080-04768d870339" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1481.870192] env[65522]: DEBUG nova.network.neutron [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Refreshing network info cache for port e817d4b6-77e2-43b4-932b-9ba3ecd832ef {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1482.184580] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115273, 'name': CreateVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1482.216369] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 9f342b89-bde2-4c35-ae42-cfe1e6973b74] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1482.372893] env[65522]: WARNING neutronclient.v2_0.client [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1482.373586] env[65522]: WARNING openstack [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1482.373938] env[65522]: WARNING openstack [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1482.465205] env[65522]: WARNING openstack [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1482.465586] env[65522]: WARNING openstack [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1482.522152] env[65522]: WARNING neutronclient.v2_0.client [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1482.522834] env[65522]: WARNING openstack [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1482.523201] env[65522]: WARNING openstack [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1482.601661] env[65522]: DEBUG nova.network.neutron [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Updated VIF entry in instance network info cache for port e817d4b6-77e2-43b4-932b-9ba3ecd832ef. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1482.602046] env[65522]: DEBUG nova.network.neutron [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Updating instance_info_cache with network_info: [{"id": "e817d4b6-77e2-43b4-932b-9ba3ecd832ef", "address": "fa:16:3e:1b:28:32", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape817d4b6-77", "ovs_interfaceid": "e817d4b6-77e2-43b4-932b-9ba3ecd832ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1482.684560] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115273, 'name': CreateVM_Task, 'duration_secs': 0.556293} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1482.684736] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1482.685263] env[65522]: WARNING neutronclient.v2_0.client [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1482.685651] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1482.685810] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1482.686151] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1482.686440] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ad00af0-9517-4571-9668-90822929d6f0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1482.691547] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1482.691547] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521e303d-7d5e-1108-8022-bc20cd5019e3" [ 1482.691547] env[65522]: _type = "Task" [ 1482.691547] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1482.700477] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521e303d-7d5e-1108-8022-bc20cd5019e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1482.720065] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 4aca4e87-b923-49e4-88d8-7ff51bcbe993] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1483.105138] env[65522]: DEBUG oslo_concurrency.lockutils [req-6975d66b-c73b-4357-b5de-41e361df8017 req-d659e8ac-e440-4047-b0db-aa58dffcc74d service nova] Releasing lock "refresh_cache-2ad40e32-4111-4f49-9080-04768d870339" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1483.202768] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]521e303d-7d5e-1108-8022-bc20cd5019e3, 'name': SearchDatastore_Task, 'duration_secs': 0.01308} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1483.203072] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1483.203316] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1483.203553] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1483.203694] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1483.203875] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1483.204157] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-05fcbe3e-aef8-4917-8443-22f9eb84f4a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.213359] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1483.213535] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1483.214261] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-002ec709-aeff-4401-9fc3-6ee162883fd8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.219533] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1483.219533] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52354024-d986-1296-82c8-5be5bcce8fc2" [ 1483.219533] env[65522]: _type = "Task" [ 1483.219533] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1483.222952] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] [instance: 7617189c-a902-42e7-8165-0e7c4a1de06d] Instance has had 0 of 5 cleanup attempts {{(pid=65522) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11946}} [ 1483.229668] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52354024-d986-1296-82c8-5be5bcce8fc2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1483.730300] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52354024-d986-1296-82c8-5be5bcce8fc2, 'name': SearchDatastore_Task, 'duration_secs': 0.010032} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1483.731028] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75199eab-03d2-4fd0-8883-4c80c7cc5149 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.736151] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1483.736151] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a12930-bd65-e4f6-53cc-c5dbcf62096e" [ 1483.736151] env[65522]: _type = "Task" [ 1483.736151] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1483.744350] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a12930-bd65-e4f6-53cc-c5dbcf62096e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1484.247485] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52a12930-bd65-e4f6-53cc-c5dbcf62096e, 'name': SearchDatastore_Task, 'duration_secs': 0.010699} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1484.247909] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1484.248047] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 2ad40e32-4111-4f49-9080-04768d870339/2ad40e32-4111-4f49-9080-04768d870339.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1484.248327] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-70105f36-a5fb-4789-9d91-19f15f135718 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.255575] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1484.255575] env[65522]: value = "task-5115274" [ 1484.255575] env[65522]: _type = "Task" [ 1484.255575] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1484.263778] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115274, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1484.766122] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115274, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439593} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1484.766332] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 2ad40e32-4111-4f49-9080-04768d870339/2ad40e32-4111-4f49-9080-04768d870339.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1484.766567] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1484.766911] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-60ec3802-8c9a-408e-8914-aa10bb1d8864 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.775662] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1484.775662] env[65522]: value = "task-5115275" [ 1484.775662] env[65522]: _type = "Task" [ 1484.775662] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1484.783698] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115275, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1485.285668] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115275, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066796} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1485.286148] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1485.286776] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35a45fc-7005-4aa4-a724-ec4c3346a965 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1485.310177] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Reconfiguring VM instance instance-0000007d to attach disk [datastore1] 2ad40e32-4111-4f49-9080-04768d870339/2ad40e32-4111-4f49-9080-04768d870339.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1485.310486] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6838a34b-d95f-4a21-a811-35aecfc1083c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1485.330648] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1485.330648] env[65522]: value = "task-5115276" [ 1485.330648] env[65522]: _type = "Task" [ 1485.330648] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1485.338990] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115276, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1485.840708] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115276, 'name': ReconfigVM_Task, 'duration_secs': 0.310211} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1485.840994] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Reconfigured VM instance instance-0000007d to attach disk [datastore1] 2ad40e32-4111-4f49-9080-04768d870339/2ad40e32-4111-4f49-9080-04768d870339.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1485.841711] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf1ec1d3-bc81-4ea7-8175-00c3b097ab0b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1485.849248] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1485.849248] env[65522]: value = "task-5115277" [ 1485.849248] env[65522]: _type = "Task" [ 1485.849248] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1485.858022] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115277, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1486.359300] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115277, 'name': Rename_Task, 'duration_secs': 0.150141} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1486.359696] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1486.359812] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8973eed1-784c-4166-b869-06cb5f7bf5e0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.367024] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1486.367024] env[65522]: value = "task-5115278" [ 1486.367024] env[65522]: _type = "Task" [ 1486.367024] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1486.375308] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115278, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1486.877765] env[65522]: DEBUG oslo_vmware.api [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115278, 'name': PowerOnVM_Task, 'duration_secs': 0.465206} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1486.877985] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1486.878206] env[65522]: INFO nova.compute.manager [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Took 7.31 seconds to spawn the instance on the hypervisor. [ 1486.878384] env[65522]: DEBUG nova.compute.manager [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1486.879226] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2d234e-9de7-4c65-aa95-f5fbcc47813a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.399980] env[65522]: INFO nova.compute.manager [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Took 11.99 seconds to build instance. [ 1487.902378] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1e8feefb-a055-4e8b-b0a0-ec94ee47d61e tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "2ad40e32-4111-4f49-9080-04768d870339" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.503s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1488.542042] env[65522]: DEBUG nova.compute.manager [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Received event network-changed-e817d4b6-77e2-43b4-932b-9ba3ecd832ef {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1488.542384] env[65522]: DEBUG nova.compute.manager [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Refreshing instance network info cache due to event network-changed-e817d4b6-77e2-43b4-932b-9ba3ecd832ef. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1488.542384] env[65522]: DEBUG oslo_concurrency.lockutils [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] Acquiring lock "refresh_cache-2ad40e32-4111-4f49-9080-04768d870339" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1488.542535] env[65522]: DEBUG oslo_concurrency.lockutils [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] Acquired lock "refresh_cache-2ad40e32-4111-4f49-9080-04768d870339" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1488.542709] env[65522]: DEBUG nova.network.neutron [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Refreshing network info cache for port e817d4b6-77e2-43b4-932b-9ba3ecd832ef {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1489.046207] env[65522]: WARNING neutronclient.v2_0.client [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1489.046943] env[65522]: WARNING openstack [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1489.047417] env[65522]: WARNING openstack [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1489.161483] env[65522]: WARNING openstack [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1489.161867] env[65522]: WARNING openstack [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1489.220276] env[65522]: WARNING neutronclient.v2_0.client [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1489.220920] env[65522]: WARNING openstack [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1489.221329] env[65522]: WARNING openstack [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1489.301894] env[65522]: DEBUG nova.network.neutron [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Updated VIF entry in instance network info cache for port e817d4b6-77e2-43b4-932b-9ba3ecd832ef. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1489.302404] env[65522]: DEBUG nova.network.neutron [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Updating instance_info_cache with network_info: [{"id": "e817d4b6-77e2-43b4-932b-9ba3ecd832ef", "address": "fa:16:3e:1b:28:32", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape817d4b6-77", "ovs_interfaceid": "e817d4b6-77e2-43b4-932b-9ba3ecd832ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1489.805852] env[65522]: DEBUG oslo_concurrency.lockutils [req-6c98b59f-1007-4a14-aed4-5bad3be0ec98 req-77c9f46c-5afe-4bc6-a93f-5b29752271be service nova] Releasing lock "refresh_cache-2ad40e32-4111-4f49-9080-04768d870339" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1524.677999] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "2ad40e32-4111-4f49-9080-04768d870339" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1524.678515] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "2ad40e32-4111-4f49-9080-04768d870339" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1524.678593] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "2ad40e32-4111-4f49-9080-04768d870339-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1524.678860] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "2ad40e32-4111-4f49-9080-04768d870339-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1524.679172] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "2ad40e32-4111-4f49-9080-04768d870339-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1524.682843] env[65522]: INFO nova.compute.manager [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Terminating instance [ 1525.186840] env[65522]: DEBUG nova.compute.manager [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1525.187063] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1525.188053] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fdbcd4d-8500-4cc5-b4e4-3f4811198bd9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1525.196638] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1525.196927] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67d5d90c-5a75-4533-a285-0f58b7ad5af7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1525.203696] env[65522]: DEBUG oslo_vmware.api [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1525.203696] env[65522]: value = "task-5115279" [ 1525.203696] env[65522]: _type = "Task" [ 1525.203696] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1525.212837] env[65522]: DEBUG oslo_vmware.api [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115279, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1525.717332] env[65522]: DEBUG oslo_vmware.api [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115279, 'name': PowerOffVM_Task, 'duration_secs': 0.190318} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1525.717720] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1525.717930] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1525.718331] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-18c926f2-e1ac-48c6-ac1d-d3dd54923183 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1525.795929] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1525.796197] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1525.796386] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Deleting the datastore file [datastore1] 2ad40e32-4111-4f49-9080-04768d870339 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1525.796715] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a88fca1-e26a-43da-a201-8afcfe6f40f5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1525.805598] env[65522]: DEBUG oslo_vmware.api [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1525.805598] env[65522]: value = "task-5115281" [ 1525.805598] env[65522]: _type = "Task" [ 1525.805598] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1525.814088] env[65522]: DEBUG oslo_vmware.api [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115281, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1526.315649] env[65522]: DEBUG oslo_vmware.api [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115281, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140931} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1526.315929] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1526.316127] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1526.316318] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1526.316488] env[65522]: INFO nova.compute.manager [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1526.316806] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1526.317022] env[65522]: DEBUG nova.compute.manager [-] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1526.317125] env[65522]: DEBUG nova.network.neutron [-] [instance: 2ad40e32-4111-4f49-9080-04768d870339] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1526.317372] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1526.317911] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1526.318175] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1526.377362] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1526.854115] env[65522]: DEBUG nova.compute.manager [req-00cc6c7e-373e-4f19-9fa7-1df34d49fd24 req-20dc91da-ee9a-4f70-8a7d-e1b5b095cb2d service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Received event network-vif-deleted-e817d4b6-77e2-43b4-932b-9ba3ecd832ef {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1526.854115] env[65522]: INFO nova.compute.manager [req-00cc6c7e-373e-4f19-9fa7-1df34d49fd24 req-20dc91da-ee9a-4f70-8a7d-e1b5b095cb2d service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Neutron deleted interface e817d4b6-77e2-43b4-932b-9ba3ecd832ef; detaching it from the instance and deleting it from the info cache [ 1526.854115] env[65522]: DEBUG nova.network.neutron [req-00cc6c7e-373e-4f19-9fa7-1df34d49fd24 req-20dc91da-ee9a-4f70-8a7d-e1b5b095cb2d service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1527.334292] env[65522]: DEBUG nova.network.neutron [-] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1527.357380] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b4f3d79b-f373-443a-a88e-d079fb3f681c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1527.368211] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fc31d83-bce1-44c0-8243-042db92c5978 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1527.396891] env[65522]: DEBUG nova.compute.manager [req-00cc6c7e-373e-4f19-9fa7-1df34d49fd24 req-20dc91da-ee9a-4f70-8a7d-e1b5b095cb2d service nova] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Detach interface failed, port_id=e817d4b6-77e2-43b4-932b-9ba3ecd832ef, reason: Instance 2ad40e32-4111-4f49-9080-04768d870339 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1527.837834] env[65522]: INFO nova.compute.manager [-] [instance: 2ad40e32-4111-4f49-9080-04768d870339] Took 1.52 seconds to deallocate network for instance. [ 1528.344306] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1528.344741] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1528.344823] env[65522]: DEBUG nova.objects.instance [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lazy-loading 'resources' on Instance uuid 2ad40e32-4111-4f49-9080-04768d870339 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1528.730641] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1528.730884] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1528.890495] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac6bd54-75f0-4190-bb68-5a0742e2c129 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.899202] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9408334e-862c-41b4-a8d3-cbbb36ad28f1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.931401] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1a4404-2c26-4c6f-b5a1-f40d50797272 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.939963] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2fcf28-ed46-4fbb-a077-0a4b76f03b2f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.954456] env[65522]: DEBUG nova.compute.provider_tree [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1529.236614] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.236614] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.236614] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.236917] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.236917] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.237038] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.237725] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1529.237725] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.458426] env[65522]: DEBUG nova.scheduler.client.report [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1529.740426] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1529.963744] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1529.966135] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.226s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1529.966324] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1529.966479] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1529.967540] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9b7fdd-5efc-4d94-ac00-de7b67301a03 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1529.976638] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed9803ae-b984-466a-8281-882f33fa2280 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1529.991865] env[65522]: INFO nova.scheduler.client.report [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Deleted allocations for instance 2ad40e32-4111-4f49-9080-04768d870339 [ 1529.993792] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89233785-300a-4557-a6ce-40df4e363369 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.002531] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8327b06-8019-4825-a8f9-d1e3a41d6bb9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.034083] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180341MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1530.034878] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1530.034878] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1530.502304] env[65522]: DEBUG oslo_concurrency.lockutils [None req-6b2c6e16-dc66-44f2-b425-a5091f3ad8cf tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "2ad40e32-4111-4f49-9080-04768d870339" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.824s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1531.059335] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 148f338e-d583-46b7-8ae0-855cb280cd47 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1531.059335] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1531.059335] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=100GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '1', 'num_vm_active': '1', 'num_task_None': '1', 'num_os_type_None': '1', 'num_proj_37a72785dad54301a64694398f167f92': '1', 'io_workload': '0'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1531.087640] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a10b6678-e048-4dc6-8a1b-c1bdb37a596c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.095983] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc820444-137c-40a2-baad-714e69fdf8ec {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.128585] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-093a8d09-fcc7-4dcd-831e-9dc9e95a2e24 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.137111] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498f179d-84e7-4971-b456-46b38a8f1c42 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.151243] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1531.654611] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1531.821926] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "148f338e-d583-46b7-8ae0-855cb280cd47" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1531.822198] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "148f338e-d583-46b7-8ae0-855cb280cd47" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1532.159525] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1532.159808] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.125s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1532.325613] env[65522]: INFO nova.compute.manager [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Detaching volume 384aa72c-340c-4358-8b76-0c62fe61d3aa [ 1532.366020] env[65522]: INFO nova.virt.block_device [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Attempting to driver detach volume 384aa72c-340c-4358-8b76-0c62fe61d3aa from mountpoint /dev/sdb [ 1532.366405] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1532.366689] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995002', 'volume_id': '384aa72c-340c-4358-8b76-0c62fe61d3aa', 'name': 'volume-384aa72c-340c-4358-8b76-0c62fe61d3aa', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '148f338e-d583-46b7-8ae0-855cb280cd47', 'attached_at': '', 'detached_at': '', 'volume_id': '384aa72c-340c-4358-8b76-0c62fe61d3aa', 'serial': '384aa72c-340c-4358-8b76-0c62fe61d3aa'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1532.368079] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6a3836-a5bd-49d0-99f1-2155dad18a9e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1532.403531] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f84c67e8-03a4-4b13-833c-eae9cfb6f127 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1532.414107] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67469d8-3b58-4e81-b20f-6a6f023a4587 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1532.447385] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a378d0a-4316-4385-8d18-5c052f6a8fd8 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1532.464327] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The volume has not been displaced from its original location: [datastore1] volume-384aa72c-340c-4358-8b76-0c62fe61d3aa/volume-384aa72c-340c-4358-8b76-0c62fe61d3aa.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1532.469916] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Reconfiguring VM instance instance-0000007c to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1532.470420] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3c62203-60d1-4ac2-b5ae-34383e05cff5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1532.490292] env[65522]: DEBUG oslo_vmware.api [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1532.490292] env[65522]: value = "task-5115282" [ 1532.490292] env[65522]: _type = "Task" [ 1532.490292] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1532.499782] env[65522]: DEBUG oslo_vmware.api [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115282, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1533.000696] env[65522]: DEBUG oslo_vmware.api [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115282, 'name': ReconfigVM_Task, 'duration_secs': 0.314755} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1533.001153] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Reconfigured VM instance instance-0000007c to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1533.005760] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a6f62ca-637b-4993-a0f4-960836cb94e5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1533.023132] env[65522]: DEBUG oslo_vmware.api [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1533.023132] env[65522]: value = "task-5115283" [ 1533.023132] env[65522]: _type = "Task" [ 1533.023132] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1533.032585] env[65522]: DEBUG oslo_vmware.api [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115283, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1533.535713] env[65522]: DEBUG oslo_vmware.api [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115283, 'name': ReconfigVM_Task, 'duration_secs': 0.167232} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1533.535963] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995002', 'volume_id': '384aa72c-340c-4358-8b76-0c62fe61d3aa', 'name': 'volume-384aa72c-340c-4358-8b76-0c62fe61d3aa', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '148f338e-d583-46b7-8ae0-855cb280cd47', 'attached_at': '', 'detached_at': '', 'volume_id': '384aa72c-340c-4358-8b76-0c62fe61d3aa', 'serial': '384aa72c-340c-4358-8b76-0c62fe61d3aa'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1534.080130] env[65522]: DEBUG nova.objects.instance [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lazy-loading 'flavor' on Instance uuid 148f338e-d583-46b7-8ae0-855cb280cd47 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1535.088547] env[65522]: DEBUG oslo_concurrency.lockutils [None req-8dc69e20-da38-4af3-b1cd-781a87e8cef9 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "148f338e-d583-46b7-8ae0-855cb280cd47" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.266s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1536.087144] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "148f338e-d583-46b7-8ae0-855cb280cd47" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1536.087417] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "148f338e-d583-46b7-8ae0-855cb280cd47" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1536.087636] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "148f338e-d583-46b7-8ae0-855cb280cd47-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1536.087828] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "148f338e-d583-46b7-8ae0-855cb280cd47-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1536.087993] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "148f338e-d583-46b7-8ae0-855cb280cd47-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1536.090413] env[65522]: INFO nova.compute.manager [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Terminating instance [ 1536.594378] env[65522]: DEBUG nova.compute.manager [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1536.594651] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1536.595773] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11505602-e3e1-4c48-9d5f-a468304209eb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1536.604982] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1536.605285] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7da309cb-e524-4100-a89f-335e111ba22c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1536.612513] env[65522]: DEBUG oslo_vmware.api [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1536.612513] env[65522]: value = "task-5115284" [ 1536.612513] env[65522]: _type = "Task" [ 1536.612513] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1536.621943] env[65522]: DEBUG oslo_vmware.api [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115284, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1537.122287] env[65522]: DEBUG oslo_vmware.api [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115284, 'name': PowerOffVM_Task, 'duration_secs': 0.201165} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1537.122685] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1537.122685] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1537.122915] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e72e2db-702e-44ab-aa07-850819e11b95 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1537.187153] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1537.187383] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1537.187566] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Deleting the datastore file [datastore1] 148f338e-d583-46b7-8ae0-855cb280cd47 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1537.187867] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-84f2d240-b8ed-4133-8309-f5a8f229e23a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1537.198372] env[65522]: DEBUG oslo_vmware.api [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1537.198372] env[65522]: value = "task-5115286" [ 1537.198372] env[65522]: _type = "Task" [ 1537.198372] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1537.207086] env[65522]: DEBUG oslo_vmware.api [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115286, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1537.709109] env[65522]: DEBUG oslo_vmware.api [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115286, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158177} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1537.709374] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1537.709549] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1537.709728] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1537.709957] env[65522]: INFO nova.compute.manager [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1537.710259] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1537.710469] env[65522]: DEBUG nova.compute.manager [-] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1537.710567] env[65522]: DEBUG nova.network.neutron [-] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1537.710806] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1537.711364] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1537.711620] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1537.748257] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1538.213024] env[65522]: DEBUG nova.compute.manager [req-12a91a04-7416-44cd-83ad-818413a4672f req-5c76daeb-bb4e-4a3d-a4f7-1cd81970b617 service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Received event network-vif-deleted-df1f3178-7f92-496b-8bc7-908acf23200f {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1538.213342] env[65522]: INFO nova.compute.manager [req-12a91a04-7416-44cd-83ad-818413a4672f req-5c76daeb-bb4e-4a3d-a4f7-1cd81970b617 service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Neutron deleted interface df1f3178-7f92-496b-8bc7-908acf23200f; detaching it from the instance and deleting it from the info cache [ 1538.213669] env[65522]: DEBUG nova.network.neutron [req-12a91a04-7416-44cd-83ad-818413a4672f req-5c76daeb-bb4e-4a3d-a4f7-1cd81970b617 service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1538.695085] env[65522]: DEBUG nova.network.neutron [-] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1538.717053] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c2b633a-c382-4e42-86b7-2d30e8cdd403 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.727265] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3764cb-adbb-45e9-85b3-938856b201e4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.754626] env[65522]: DEBUG nova.compute.manager [req-12a91a04-7416-44cd-83ad-818413a4672f req-5c76daeb-bb4e-4a3d-a4f7-1cd81970b617 service nova] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Detach interface failed, port_id=df1f3178-7f92-496b-8bc7-908acf23200f, reason: Instance 148f338e-d583-46b7-8ae0-855cb280cd47 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1539.201404] env[65522]: INFO nova.compute.manager [-] [instance: 148f338e-d583-46b7-8ae0-855cb280cd47] Took 1.49 seconds to deallocate network for instance. [ 1539.708125] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1539.708545] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1539.708598] env[65522]: DEBUG nova.objects.instance [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lazy-loading 'resources' on Instance uuid 148f338e-d583-46b7-8ae0-855cb280cd47 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1540.243535] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451a78a9-706f-4dd2-8dc1-0ac0071f9809 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.251353] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc1a067-7c2f-45b0-9a1f-8e61af582cc2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.282307] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b048487-dfdc-46b4-8f7d-e884f50fc963 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.290537] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c69c3f3-f3c2-4e37-862a-7ad27d44b603 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.307436] env[65522]: DEBUG nova.compute.provider_tree [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1540.810910] env[65522]: DEBUG nova.scheduler.client.report [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1541.315972] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.607s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1541.337904] env[65522]: INFO nova.scheduler.client.report [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Deleted allocations for instance 148f338e-d583-46b7-8ae0-855cb280cd47 [ 1541.846492] env[65522]: DEBUG oslo_concurrency.lockutils [None req-5669f8f6-6dd4-4ed1-be80-8db0be0df3ea tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "148f338e-d583-46b7-8ae0-855cb280cd47" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.759s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1544.233114] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1544.233504] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1544.735588] env[65522]: DEBUG nova.compute.manager [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1545.256131] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1545.256429] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1545.258040] env[65522]: INFO nova.compute.claims [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1546.294185] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af4480bb-37cf-4d42-bf02-e1875d6327f2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1546.302095] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d51baf-8741-4924-b726-aaa68c4a1f31 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1546.332707] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b9506f4-ddb5-41d6-8f3d-2f25f0521256 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1546.340082] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d506b1-c6eb-4cb9-b409-496a82fe9491 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1546.353146] env[65522]: DEBUG nova.compute.provider_tree [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1546.856291] env[65522]: DEBUG nova.scheduler.client.report [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1547.361303] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.105s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1547.362029] env[65522]: DEBUG nova.compute.manager [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1547.867492] env[65522]: DEBUG nova.compute.utils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1547.868944] env[65522]: DEBUG nova.compute.manager [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1547.869154] env[65522]: DEBUG nova.network.neutron [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1547.869502] env[65522]: WARNING neutronclient.v2_0.client [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1547.869812] env[65522]: WARNING neutronclient.v2_0.client [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1547.870410] env[65522]: WARNING openstack [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1547.870749] env[65522]: WARNING openstack [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1547.948716] env[65522]: DEBUG nova.policy [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f35b3e2725f49e6bd84e0aa787c0e43', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37a72785dad54301a64694398f167f92', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1548.239946] env[65522]: DEBUG nova.network.neutron [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Successfully created port: f8ea738f-3a80-44d2-8bd7-23550ddf6c35 {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1548.379582] env[65522]: DEBUG nova.compute.manager [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1549.391660] env[65522]: DEBUG nova.compute.manager [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1549.417704] env[65522]: DEBUG nova.virt.hardware [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1549.417968] env[65522]: DEBUG nova.virt.hardware [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1549.418413] env[65522]: DEBUG nova.virt.hardware [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1549.418413] env[65522]: DEBUG nova.virt.hardware [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1549.418532] env[65522]: DEBUG nova.virt.hardware [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1549.418674] env[65522]: DEBUG nova.virt.hardware [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1549.418876] env[65522]: DEBUG nova.virt.hardware [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1549.419042] env[65522]: DEBUG nova.virt.hardware [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1549.419316] env[65522]: DEBUG nova.virt.hardware [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1549.419446] env[65522]: DEBUG nova.virt.hardware [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1549.419989] env[65522]: DEBUG nova.virt.hardware [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1549.420541] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b98e87e-bf4d-44cf-892d-4050b852b6ff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1549.429232] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe29507d-27c6-4ba8-b161-6957a9ccca82 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1550.224670] env[65522]: DEBUG nova.compute.manager [req-00d3ee27-e51a-4345-850f-8006e2222386 req-2424740a-7c9a-41df-9a6c-7c73c86e2afb service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Received event network-vif-plugged-f8ea738f-3a80-44d2-8bd7-23550ddf6c35 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1550.224900] env[65522]: DEBUG oslo_concurrency.lockutils [req-00d3ee27-e51a-4345-850f-8006e2222386 req-2424740a-7c9a-41df-9a6c-7c73c86e2afb service nova] Acquiring lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1550.225157] env[65522]: DEBUG oslo_concurrency.lockutils [req-00d3ee27-e51a-4345-850f-8006e2222386 req-2424740a-7c9a-41df-9a6c-7c73c86e2afb service nova] Lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1550.225337] env[65522]: DEBUG oslo_concurrency.lockutils [req-00d3ee27-e51a-4345-850f-8006e2222386 req-2424740a-7c9a-41df-9a6c-7c73c86e2afb service nova] Lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1550.225505] env[65522]: DEBUG nova.compute.manager [req-00d3ee27-e51a-4345-850f-8006e2222386 req-2424740a-7c9a-41df-9a6c-7c73c86e2afb service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] No waiting events found dispatching network-vif-plugged-f8ea738f-3a80-44d2-8bd7-23550ddf6c35 {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1550.225665] env[65522]: WARNING nova.compute.manager [req-00d3ee27-e51a-4345-850f-8006e2222386 req-2424740a-7c9a-41df-9a6c-7c73c86e2afb service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Received unexpected event network-vif-plugged-f8ea738f-3a80-44d2-8bd7-23550ddf6c35 for instance with vm_state building and task_state spawning. [ 1550.595996] env[65522]: DEBUG nova.network.neutron [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Successfully updated port: f8ea738f-3a80-44d2-8bd7-23550ddf6c35 {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1550.640899] env[65522]: DEBUG nova.compute.manager [req-e8879c2f-f410-4069-82c4-08e2a2914fd8 req-671097e9-7e93-46f2-a44d-2a41f2d03a0c service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Received event network-changed-f8ea738f-3a80-44d2-8bd7-23550ddf6c35 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1550.641619] env[65522]: DEBUG nova.compute.manager [req-e8879c2f-f410-4069-82c4-08e2a2914fd8 req-671097e9-7e93-46f2-a44d-2a41f2d03a0c service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Refreshing instance network info cache due to event network-changed-f8ea738f-3a80-44d2-8bd7-23550ddf6c35. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1550.641619] env[65522]: DEBUG oslo_concurrency.lockutils [req-e8879c2f-f410-4069-82c4-08e2a2914fd8 req-671097e9-7e93-46f2-a44d-2a41f2d03a0c service nova] Acquiring lock "refresh_cache-fa65d148-a3fb-4d06-b1a6-aa3280c8794f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1550.641619] env[65522]: DEBUG oslo_concurrency.lockutils [req-e8879c2f-f410-4069-82c4-08e2a2914fd8 req-671097e9-7e93-46f2-a44d-2a41f2d03a0c service nova] Acquired lock "refresh_cache-fa65d148-a3fb-4d06-b1a6-aa3280c8794f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1550.641619] env[65522]: DEBUG nova.network.neutron [req-e8879c2f-f410-4069-82c4-08e2a2914fd8 req-671097e9-7e93-46f2-a44d-2a41f2d03a0c service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Refreshing network info cache for port f8ea738f-3a80-44d2-8bd7-23550ddf6c35 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1551.097914] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "refresh_cache-fa65d148-a3fb-4d06-b1a6-aa3280c8794f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1551.144762] env[65522]: WARNING neutronclient.v2_0.client [req-e8879c2f-f410-4069-82c4-08e2a2914fd8 req-671097e9-7e93-46f2-a44d-2a41f2d03a0c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1551.145555] env[65522]: WARNING openstack [req-e8879c2f-f410-4069-82c4-08e2a2914fd8 req-671097e9-7e93-46f2-a44d-2a41f2d03a0c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1551.145904] env[65522]: WARNING openstack [req-e8879c2f-f410-4069-82c4-08e2a2914fd8 req-671097e9-7e93-46f2-a44d-2a41f2d03a0c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1551.183882] env[65522]: DEBUG nova.network.neutron [req-e8879c2f-f410-4069-82c4-08e2a2914fd8 req-671097e9-7e93-46f2-a44d-2a41f2d03a0c service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1551.269208] env[65522]: DEBUG nova.network.neutron [req-e8879c2f-f410-4069-82c4-08e2a2914fd8 req-671097e9-7e93-46f2-a44d-2a41f2d03a0c service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1551.776019] env[65522]: DEBUG oslo_concurrency.lockutils [req-e8879c2f-f410-4069-82c4-08e2a2914fd8 req-671097e9-7e93-46f2-a44d-2a41f2d03a0c service nova] Releasing lock "refresh_cache-fa65d148-a3fb-4d06-b1a6-aa3280c8794f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1551.776473] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired lock "refresh_cache-fa65d148-a3fb-4d06-b1a6-aa3280c8794f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1551.776668] env[65522]: DEBUG nova.network.neutron [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1552.280089] env[65522]: WARNING openstack [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1552.280510] env[65522]: WARNING openstack [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1552.326747] env[65522]: DEBUG nova.network.neutron [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1552.345820] env[65522]: WARNING openstack [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1552.346447] env[65522]: WARNING openstack [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1552.403279] env[65522]: WARNING neutronclient.v2_0.client [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1552.403924] env[65522]: WARNING openstack [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1552.404282] env[65522]: WARNING openstack [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1552.482008] env[65522]: DEBUG nova.network.neutron [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Updating instance_info_cache with network_info: [{"id": "f8ea738f-3a80-44d2-8bd7-23550ddf6c35", "address": "fa:16:3e:27:1e:82", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8ea738f-3a", "ovs_interfaceid": "f8ea738f-3a80-44d2-8bd7-23550ddf6c35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1552.984840] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Releasing lock "refresh_cache-fa65d148-a3fb-4d06-b1a6-aa3280c8794f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1552.985308] env[65522]: DEBUG nova.compute.manager [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Instance network_info: |[{"id": "f8ea738f-3a80-44d2-8bd7-23550ddf6c35", "address": "fa:16:3e:27:1e:82", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8ea738f-3a", "ovs_interfaceid": "f8ea738f-3a80-44d2-8bd7-23550ddf6c35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1552.985777] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:1e:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '07e9bef1-2b0e-4e4d-997f-de71bb0e213a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f8ea738f-3a80-44d2-8bd7-23550ddf6c35', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1552.993284] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1552.993521] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1552.993796] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-57ef88a0-6bd5-4514-b0e0-cce7ff29f6f0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1553.014886] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1553.014886] env[65522]: value = "task-5115287" [ 1553.014886] env[65522]: _type = "Task" [ 1553.014886] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1553.023680] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115287, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1553.525804] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115287, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1554.027817] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115287, 'name': CreateVM_Task} progress is 25%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1554.527989] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115287, 'name': CreateVM_Task, 'duration_secs': 1.189965} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1554.528214] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1554.528739] env[65522]: WARNING neutronclient.v2_0.client [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1554.529124] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1554.529283] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1554.529595] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1554.529848] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d317c956-6d52-4473-a5c2-80128601b190 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1554.534547] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1554.534547] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c1ebe5-dbfc-b26f-b93a-ce3646893a32" [ 1554.534547] env[65522]: _type = "Task" [ 1554.534547] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1554.543605] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c1ebe5-dbfc-b26f-b93a-ce3646893a32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1555.045172] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c1ebe5-dbfc-b26f-b93a-ce3646893a32, 'name': SearchDatastore_Task, 'duration_secs': 0.011897} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1555.045579] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1555.045707] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1555.045952] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1555.046128] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1555.046296] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1555.046559] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c999efd5-8c99-4318-8bd5-03cf4bbb2014 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1555.056113] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1555.056304] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1555.057053] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc685323-f5d3-4c17-963e-f6734c0228f1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1555.062856] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1555.062856] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529485f8-a157-1146-438d-6118e417ec7a" [ 1555.062856] env[65522]: _type = "Task" [ 1555.062856] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1555.070522] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529485f8-a157-1146-438d-6118e417ec7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1555.575375] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]529485f8-a157-1146-438d-6118e417ec7a, 'name': SearchDatastore_Task, 'duration_secs': 0.010137} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1555.576175] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0736631d-244d-46fe-acea-d6c400a2c36c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1555.581589] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1555.581589] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525b7f03-bc36-a766-1889-86a8b13cb0d8" [ 1555.581589] env[65522]: _type = "Task" [ 1555.581589] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1555.589443] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525b7f03-bc36-a766-1889-86a8b13cb0d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1556.092254] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]525b7f03-bc36-a766-1889-86a8b13cb0d8, 'name': SearchDatastore_Task, 'duration_secs': 0.010164} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1556.092679] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1556.092769] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] fa65d148-a3fb-4d06-b1a6-aa3280c8794f/fa65d148-a3fb-4d06-b1a6-aa3280c8794f.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1556.093047] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9852e403-4ff6-4035-bf9e-e1638505e489 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1556.100556] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1556.100556] env[65522]: value = "task-5115288" [ 1556.100556] env[65522]: _type = "Task" [ 1556.100556] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1556.108670] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115288, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1556.610287] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115288, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435727} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1556.610543] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] fa65d148-a3fb-4d06-b1a6-aa3280c8794f/fa65d148-a3fb-4d06-b1a6-aa3280c8794f.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1556.610753] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1556.611013] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7c3fb50f-31f9-4001-b193-c9897618c222 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1556.618290] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1556.618290] env[65522]: value = "task-5115289" [ 1556.618290] env[65522]: _type = "Task" [ 1556.618290] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1556.626771] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115289, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1557.129798] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115289, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.185686} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1557.130256] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1557.130931] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02795a6-ed1c-41b9-bb6b-da215cf65a45 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1557.153654] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Reconfiguring VM instance instance-0000007e to attach disk [datastore1] fa65d148-a3fb-4d06-b1a6-aa3280c8794f/fa65d148-a3fb-4d06-b1a6-aa3280c8794f.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1557.153944] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90f211f6-1106-428e-9bbb-5ade79cc32bd {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1557.173196] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1557.173196] env[65522]: value = "task-5115290" [ 1557.173196] env[65522]: _type = "Task" [ 1557.173196] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1557.181376] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115290, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1557.684167] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115290, 'name': ReconfigVM_Task, 'duration_secs': 0.298162} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1557.684655] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Reconfigured VM instance instance-0000007e to attach disk [datastore1] fa65d148-a3fb-4d06-b1a6-aa3280c8794f/fa65d148-a3fb-4d06-b1a6-aa3280c8794f.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1557.685425] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-13b1a5ef-df7b-4052-b489-b48bb1d56af0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1557.692559] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1557.692559] env[65522]: value = "task-5115291" [ 1557.692559] env[65522]: _type = "Task" [ 1557.692559] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1557.702301] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115291, 'name': Rename_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1558.204122] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115291, 'name': Rename_Task, 'duration_secs': 0.142471} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1558.204122] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1558.204543] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-78b2552d-ea57-46c5-8ec7-88d610d62dcb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1558.211286] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1558.211286] env[65522]: value = "task-5115292" [ 1558.211286] env[65522]: _type = "Task" [ 1558.211286] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1558.219012] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115292, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1558.721872] env[65522]: DEBUG oslo_vmware.api [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115292, 'name': PowerOnVM_Task, 'duration_secs': 0.475538} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1558.722144] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1558.722354] env[65522]: INFO nova.compute.manager [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Took 9.33 seconds to spawn the instance on the hypervisor. [ 1558.722536] env[65522]: DEBUG nova.compute.manager [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1558.723339] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82822ebd-b425-4e3e-ab60-5b844f14e32e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1559.243016] env[65522]: INFO nova.compute.manager [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Took 14.00 seconds to build instance. [ 1559.517541] env[65522]: DEBUG nova.compute.manager [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Received event network-changed-f8ea738f-3a80-44d2-8bd7-23550ddf6c35 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1559.517807] env[65522]: DEBUG nova.compute.manager [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Refreshing instance network info cache due to event network-changed-f8ea738f-3a80-44d2-8bd7-23550ddf6c35. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1559.518320] env[65522]: DEBUG oslo_concurrency.lockutils [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] Acquiring lock "refresh_cache-fa65d148-a3fb-4d06-b1a6-aa3280c8794f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1559.518411] env[65522]: DEBUG oslo_concurrency.lockutils [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] Acquired lock "refresh_cache-fa65d148-a3fb-4d06-b1a6-aa3280c8794f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1559.518559] env[65522]: DEBUG nova.network.neutron [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Refreshing network info cache for port f8ea738f-3a80-44d2-8bd7-23550ddf6c35 {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1559.745774] env[65522]: DEBUG oslo_concurrency.lockutils [None req-9e0480d3-6c22-422c-b721-6276b16074ae tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.512s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1560.022048] env[65522]: WARNING neutronclient.v2_0.client [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1560.022611] env[65522]: WARNING openstack [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1560.022956] env[65522]: WARNING openstack [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1560.138673] env[65522]: WARNING openstack [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1560.139113] env[65522]: WARNING openstack [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1560.197996] env[65522]: WARNING neutronclient.v2_0.client [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1560.198630] env[65522]: WARNING openstack [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1560.199022] env[65522]: WARNING openstack [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1560.278929] env[65522]: DEBUG nova.network.neutron [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Updated VIF entry in instance network info cache for port f8ea738f-3a80-44d2-8bd7-23550ddf6c35. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1560.279322] env[65522]: DEBUG nova.network.neutron [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Updating instance_info_cache with network_info: [{"id": "f8ea738f-3a80-44d2-8bd7-23550ddf6c35", "address": "fa:16:3e:27:1e:82", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf8ea738f-3a", "ovs_interfaceid": "f8ea738f-3a80-44d2-8bd7-23550ddf6c35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1560.782596] env[65522]: DEBUG oslo_concurrency.lockutils [req-852dc8bc-68d3-4ed3-a729-5cf36eb71e25 req-78c2a193-0000-4359-b3d3-e3a261eecff5 service nova] Releasing lock "refresh_cache-fa65d148-a3fb-4d06-b1a6-aa3280c8794f" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1592.161756] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1592.162231] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1592.162231] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1592.162384] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1592.162534] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1592.162675] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1592.162817] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1592.162953] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1592.163106] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1592.666333] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1592.667515] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1592.667515] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1592.667515] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1592.667844] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5cbf96-df13-46b7-ac0a-7df45746b88f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1592.676534] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bcdb3e5-b8df-4e44-93ff-be008a9e5a7b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1592.693667] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f89bad0-3f96-47fd-8469-caa66083b5bb {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1592.701522] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad78adc-937d-4592-957b-02fd6568e61e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1592.731222] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180326MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1592.731416] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1592.731603] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1593.758605] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance fa65d148-a3fb-4d06-b1a6-aa3280c8794f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1593.758940] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1593.758940] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=100GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '1', 'num_vm_active': '1', 'num_task_None': '1', 'num_os_type_None': '1', 'num_proj_37a72785dad54301a64694398f167f92': '1', 'io_workload': '0'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1593.775396] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Refreshing inventories for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1593.787768] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Updating ProviderTree inventory for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1593.787952] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Updating inventory in ProviderTree for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1593.798283] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Refreshing aggregate associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, aggregates: None {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1593.817173] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Refreshing trait associations for resource provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d, traits: COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=65522) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1593.841184] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc50ef43-4c66-46de-97a8-a42009e9ae1e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.848672] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c686683-dc72-4067-8bcd-2f6e9c8256a7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.879133] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f47ae2-5cc1-43cc-b7ec-0109ea36d3a3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.886579] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfec092d-84bd-4b92-9ce5-ca84852e3f3d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.899861] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1594.403244] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1594.908822] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1594.909238] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.177s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1595.734667] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1595.734903] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1596.237924] env[65522]: DEBUG nova.compute.utils [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1596.741671] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1597.802511] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1597.802971] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1597.802971] env[65522]: INFO nova.compute.manager [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Attaching volume e4dec57e-6de5-4ce5-9370-112293fcd880 to /dev/sdb [ 1597.833591] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfef531-5b56-4326-ba86-73c6b8210206 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1597.841016] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37f367c-c0a3-4ff6-9e4c-3f76bfef5699 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1597.854339] env[65522]: DEBUG nova.virt.block_device [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Updating existing volume attachment record: 0017036c-1cc0-4569-b2e7-47f68afad72a {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1602.399312] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1602.399573] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995005', 'volume_id': 'e4dec57e-6de5-4ce5-9370-112293fcd880', 'name': 'volume-e4dec57e-6de5-4ce5-9370-112293fcd880', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'fa65d148-a3fb-4d06-b1a6-aa3280c8794f', 'attached_at': '', 'detached_at': '', 'volume_id': 'e4dec57e-6de5-4ce5-9370-112293fcd880', 'serial': 'e4dec57e-6de5-4ce5-9370-112293fcd880'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1602.400428] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e2bfcb-7d3e-47ed-90a6-6bddd29dc9f0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1602.418476] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50240cc2-a3a4-4fc7-aaea-518500f325d9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1602.444114] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Reconfiguring VM instance instance-0000007e to attach disk [datastore1] volume-e4dec57e-6de5-4ce5-9370-112293fcd880/volume-e4dec57e-6de5-4ce5-9370-112293fcd880.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1602.444386] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d2269b8-82e0-46d6-85d1-9d4cac7b6d2f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1602.463509] env[65522]: DEBUG oslo_vmware.api [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1602.463509] env[65522]: value = "task-5115295" [ 1602.463509] env[65522]: _type = "Task" [ 1602.463509] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1602.473899] env[65522]: DEBUG oslo_vmware.api [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115295, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1602.973535] env[65522]: DEBUG oslo_vmware.api [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115295, 'name': ReconfigVM_Task, 'duration_secs': 0.34385} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1602.974134] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Reconfigured VM instance instance-0000007e to attach disk [datastore1] volume-e4dec57e-6de5-4ce5-9370-112293fcd880/volume-e4dec57e-6de5-4ce5-9370-112293fcd880.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1602.978968] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2bb6be1d-daf1-4c7a-bccc-3b68d73cdaa1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1602.994045] env[65522]: DEBUG oslo_vmware.api [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1602.994045] env[65522]: value = "task-5115296" [ 1602.994045] env[65522]: _type = "Task" [ 1602.994045] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1603.002986] env[65522]: DEBUG oslo_vmware.api [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115296, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1603.504692] env[65522]: DEBUG oslo_vmware.api [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115296, 'name': ReconfigVM_Task, 'duration_secs': 0.146044} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1603.505068] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995005', 'volume_id': 'e4dec57e-6de5-4ce5-9370-112293fcd880', 'name': 'volume-e4dec57e-6de5-4ce5-9370-112293fcd880', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'fa65d148-a3fb-4d06-b1a6-aa3280c8794f', 'attached_at': '', 'detached_at': '', 'volume_id': 'e4dec57e-6de5-4ce5-9370-112293fcd880', 'serial': 'e4dec57e-6de5-4ce5-9370-112293fcd880'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1604.542335] env[65522]: DEBUG nova.objects.instance [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lazy-loading 'flavor' on Instance uuid fa65d148-a3fb-4d06-b1a6-aa3280c8794f {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1605.049499] env[65522]: DEBUG oslo_concurrency.lockutils [None req-b2368157-c8da-4e98-8369-8bfbecc0b764 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.247s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1605.259970] env[65522]: DEBUG oslo_concurrency.lockutils [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1605.260267] env[65522]: DEBUG oslo_concurrency.lockutils [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1605.763342] env[65522]: INFO nova.compute.manager [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Detaching volume e4dec57e-6de5-4ce5-9370-112293fcd880 [ 1605.797861] env[65522]: INFO nova.virt.block_device [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Attempting to driver detach volume e4dec57e-6de5-4ce5-9370-112293fcd880 from mountpoint /dev/sdb [ 1605.798105] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1605.798296] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995005', 'volume_id': 'e4dec57e-6de5-4ce5-9370-112293fcd880', 'name': 'volume-e4dec57e-6de5-4ce5-9370-112293fcd880', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'fa65d148-a3fb-4d06-b1a6-aa3280c8794f', 'attached_at': '', 'detached_at': '', 'volume_id': 'e4dec57e-6de5-4ce5-9370-112293fcd880', 'serial': 'e4dec57e-6de5-4ce5-9370-112293fcd880'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1605.799219] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad305b8a-14e2-49fc-8a36-fb711bce7fb3 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1605.823541] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c65607-5fce-4c02-b638-90482587b320 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1605.831455] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc692b3-f4c5-443a-8c0c-57e5b0e556f2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1605.852922] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2950b51b-25f2-4438-8669-d0467a2af392 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1605.868851] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The volume has not been displaced from its original location: [datastore1] volume-e4dec57e-6de5-4ce5-9370-112293fcd880/volume-e4dec57e-6de5-4ce5-9370-112293fcd880.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1605.874219] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Reconfiguring VM instance instance-0000007e to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1605.874586] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b98760b8-e5f8-44c7-bfde-672b8a32945d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1605.894018] env[65522]: DEBUG oslo_vmware.api [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1605.894018] env[65522]: value = "task-5115297" [ 1605.894018] env[65522]: _type = "Task" [ 1605.894018] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1605.902654] env[65522]: DEBUG oslo_vmware.api [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115297, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1606.404440] env[65522]: DEBUG oslo_vmware.api [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115297, 'name': ReconfigVM_Task, 'duration_secs': 0.231324} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1606.404656] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Reconfigured VM instance instance-0000007e to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1606.409466] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a7072a7-fc09-45f6-972c-6d16b787da3e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1606.424994] env[65522]: DEBUG oslo_vmware.api [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1606.424994] env[65522]: value = "task-5115298" [ 1606.424994] env[65522]: _type = "Task" [ 1606.424994] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1606.434030] env[65522]: DEBUG oslo_vmware.api [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115298, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1606.936872] env[65522]: DEBUG oslo_vmware.api [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115298, 'name': ReconfigVM_Task, 'duration_secs': 0.145322} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1606.937375] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995005', 'volume_id': 'e4dec57e-6de5-4ce5-9370-112293fcd880', 'name': 'volume-e4dec57e-6de5-4ce5-9370-112293fcd880', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'fa65d148-a3fb-4d06-b1a6-aa3280c8794f', 'attached_at': '', 'detached_at': '', 'volume_id': 'e4dec57e-6de5-4ce5-9370-112293fcd880', 'serial': 'e4dec57e-6de5-4ce5-9370-112293fcd880'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1607.479943] env[65522]: DEBUG nova.objects.instance [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lazy-loading 'flavor' on Instance uuid fa65d148-a3fb-4d06-b1a6-aa3280c8794f {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1608.487690] env[65522]: DEBUG oslo_concurrency.lockutils [None req-578d03e5-27f2-4981-bb65-33621049e380 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.227s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1608.811719] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1608.812035] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1608.812278] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1608.812461] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1608.812633] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1608.814922] env[65522]: INFO nova.compute.manager [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Terminating instance [ 1609.319484] env[65522]: DEBUG nova.compute.manager [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1609.319734] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1609.320953] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8c4e3a-f18d-477b-909f-caf940a59d8a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.330932] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1609.331295] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c7aa7785-67d6-40cb-a1df-3faae3c5b611 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.339197] env[65522]: DEBUG oslo_vmware.api [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1609.339197] env[65522]: value = "task-5115299" [ 1609.339197] env[65522]: _type = "Task" [ 1609.339197] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1609.351114] env[65522]: DEBUG oslo_vmware.api [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115299, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1609.849588] env[65522]: DEBUG oslo_vmware.api [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115299, 'name': PowerOffVM_Task, 'duration_secs': 0.173074} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1609.849985] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1609.850065] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1609.850287] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6af36220-2353-4b37-aeba-3d7bd1b23dda {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.919593] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1609.919841] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1609.920078] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Deleting the datastore file [datastore1] fa65d148-a3fb-4d06-b1a6-aa3280c8794f {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1609.920323] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6c3903d7-6028-4a76-a7cd-f4124e5bf0ff {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.930797] env[65522]: DEBUG oslo_vmware.api [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1609.930797] env[65522]: value = "task-5115301" [ 1609.930797] env[65522]: _type = "Task" [ 1609.930797] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1609.940893] env[65522]: DEBUG oslo_vmware.api [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115301, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1610.441343] env[65522]: DEBUG oslo_vmware.api [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115301, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158721} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1610.441589] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1610.441778] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1610.441952] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1610.442145] env[65522]: INFO nova.compute.manager [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1610.442396] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1610.442592] env[65522]: DEBUG nova.compute.manager [-] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1610.442692] env[65522]: DEBUG nova.network.neutron [-] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1610.442937] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1610.443498] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1610.443759] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1610.480685] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1610.982920] env[65522]: DEBUG nova.compute.manager [req-146126b8-db83-4ffe-bb38-3c282016f17a req-6e9f200b-59b3-4627-9540-176aa3bb325c service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Received event network-vif-deleted-f8ea738f-3a80-44d2-8bd7-23550ddf6c35 {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1610.982920] env[65522]: INFO nova.compute.manager [req-146126b8-db83-4ffe-bb38-3c282016f17a req-6e9f200b-59b3-4627-9540-176aa3bb325c service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Neutron deleted interface f8ea738f-3a80-44d2-8bd7-23550ddf6c35; detaching it from the instance and deleting it from the info cache [ 1610.982920] env[65522]: DEBUG nova.network.neutron [req-146126b8-db83-4ffe-bb38-3c282016f17a req-6e9f200b-59b3-4627-9540-176aa3bb325c service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1611.436027] env[65522]: DEBUG nova.network.neutron [-] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1611.485745] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b283a63-c0b8-43b1-9e9d-703b174053a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1611.496424] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f84cae-b75d-416e-a0f7-e665cf79c06d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1611.525666] env[65522]: DEBUG nova.compute.manager [req-146126b8-db83-4ffe-bb38-3c282016f17a req-6e9f200b-59b3-4627-9540-176aa3bb325c service nova] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Detach interface failed, port_id=f8ea738f-3a80-44d2-8bd7-23550ddf6c35, reason: Instance fa65d148-a3fb-4d06-b1a6-aa3280c8794f could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1611.938180] env[65522]: INFO nova.compute.manager [-] [instance: fa65d148-a3fb-4d06-b1a6-aa3280c8794f] Took 1.50 seconds to deallocate network for instance. [ 1612.445614] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1612.446081] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1612.446133] env[65522]: DEBUG nova.objects.instance [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lazy-loading 'resources' on Instance uuid fa65d148-a3fb-4d06-b1a6-aa3280c8794f {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1612.982326] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b4d7f4-2015-4b4f-bc8d-32bd61ff669a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1612.990660] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a17f96-a99a-4b61-88e4-0526f79921e0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1613.021675] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167665d7-2603-4649-ab72-94b460a1aeea {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1613.031195] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b824a1-95ba-4902-9ebe-a7b317193a03 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1613.046573] env[65522]: DEBUG nova.compute.provider_tree [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1613.550400] env[65522]: DEBUG nova.scheduler.client.report [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1614.058883] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.613s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1614.080788] env[65522]: INFO nova.scheduler.client.report [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Deleted allocations for instance fa65d148-a3fb-4d06-b1a6-aa3280c8794f [ 1614.591248] env[65522]: DEBUG oslo_concurrency.lockutils [None req-0cb906c4-a597-415d-87a1-b5ddad19016c tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "fa65d148-a3fb-4d06-b1a6-aa3280c8794f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.779s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1617.041073] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1617.041073] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1617.542985] env[65522]: DEBUG nova.compute.manager [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Starting instance... {{(pid=65522) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 1618.064736] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1618.065040] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1618.066672] env[65522]: INFO nova.compute.claims [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1619.104914] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c75b66f-18c4-40f3-b899-bd519484e992 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1619.113765] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb3ebff8-c57e-471a-808b-bb79667c2c35 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1619.147721] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335f2db7-8abe-43f8-8d3d-5e6bb3235c8b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1619.155991] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bffb54aa-13cf-4744-9488-24380173d574 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1619.170349] env[65522]: DEBUG nova.compute.provider_tree [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1619.674142] env[65522]: DEBUG nova.scheduler.client.report [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1620.179055] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.114s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1620.179589] env[65522]: DEBUG nova.compute.manager [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Start building networks asynchronously for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 1620.685683] env[65522]: DEBUG nova.compute.utils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1620.686478] env[65522]: DEBUG nova.compute.manager [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Allocating IP information in the background. {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 1620.686699] env[65522]: DEBUG nova.network.neutron [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] allocate_for_instance() {{(pid=65522) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1620.687120] env[65522]: WARNING neutronclient.v2_0.client [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1620.687442] env[65522]: WARNING neutronclient.v2_0.client [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1620.688131] env[65522]: WARNING openstack [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1620.688540] env[65522]: WARNING openstack [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1620.794516] env[65522]: DEBUG nova.policy [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0f35b3e2725f49e6bd84e0aa787c0e43', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37a72785dad54301a64694398f167f92', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65522) authorize /opt/stack/nova/nova/policy.py:192}} [ 1621.120214] env[65522]: DEBUG nova.network.neutron [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Successfully created port: 7c90c348-40b6-4412-9db4-4f9bf2e096dc {{(pid=65522) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1621.197420] env[65522]: DEBUG nova.compute.manager [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Start building block device mappings for instance. {{(pid=65522) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 1622.208748] env[65522]: DEBUG nova.compute.manager [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Start spawning the instance on the hypervisor. {{(pid=65522) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 1622.237638] env[65522]: DEBUG nova.virt.hardware [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-12T18:56:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-12T18:56:07Z,direct_url=,disk_format='vmdk',id=a889db67-7337-4e32-8e34-642f34402926,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='596b2b0744b64deb86a3dbe6da5c8894',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-12T18:56:08Z,virtual_size=,visibility=), allow threads: False {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1622.237966] env[65522]: DEBUG nova.virt.hardware [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Flavor limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1622.238117] env[65522]: DEBUG nova.virt.hardware [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Image limits 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1622.238283] env[65522]: DEBUG nova.virt.hardware [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Flavor pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1622.238428] env[65522]: DEBUG nova.virt.hardware [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Image pref 0:0:0 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1622.238574] env[65522]: DEBUG nova.virt.hardware [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65522) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1622.238785] env[65522]: DEBUG nova.virt.hardware [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1622.238979] env[65522]: DEBUG nova.virt.hardware [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1622.239327] env[65522]: DEBUG nova.virt.hardware [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Got 1 possible topologies {{(pid=65522) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1622.239533] env[65522]: DEBUG nova.virt.hardware [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1622.239723] env[65522]: DEBUG nova.virt.hardware [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65522) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1622.240638] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad0d13e-a40a-43dd-ab4d-038b638162c1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1622.249224] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627a4846-c510-4020-ac5a-b91d8bdcb195 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1622.525275] env[65522]: DEBUG nova.compute.manager [req-8583184a-82cd-4544-a84b-09e62186021e req-4786cb2a-32ea-4f09-b2e6-04137358d731 service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Received event network-vif-plugged-7c90c348-40b6-4412-9db4-4f9bf2e096dc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1622.526029] env[65522]: DEBUG oslo_concurrency.lockutils [req-8583184a-82cd-4544-a84b-09e62186021e req-4786cb2a-32ea-4f09-b2e6-04137358d731 service nova] Acquiring lock "96e45978-2697-4ac5-a0ba-30e95d7710f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1622.526029] env[65522]: DEBUG oslo_concurrency.lockutils [req-8583184a-82cd-4544-a84b-09e62186021e req-4786cb2a-32ea-4f09-b2e6-04137358d731 service nova] Lock "96e45978-2697-4ac5-a0ba-30e95d7710f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1622.526029] env[65522]: DEBUG oslo_concurrency.lockutils [req-8583184a-82cd-4544-a84b-09e62186021e req-4786cb2a-32ea-4f09-b2e6-04137358d731 service nova] Lock "96e45978-2697-4ac5-a0ba-30e95d7710f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1622.526029] env[65522]: DEBUG nova.compute.manager [req-8583184a-82cd-4544-a84b-09e62186021e req-4786cb2a-32ea-4f09-b2e6-04137358d731 service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] No waiting events found dispatching network-vif-plugged-7c90c348-40b6-4412-9db4-4f9bf2e096dc {{(pid=65522) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 1622.526266] env[65522]: WARNING nova.compute.manager [req-8583184a-82cd-4544-a84b-09e62186021e req-4786cb2a-32ea-4f09-b2e6-04137358d731 service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Received unexpected event network-vif-plugged-7c90c348-40b6-4412-9db4-4f9bf2e096dc for instance with vm_state building and task_state spawning. [ 1622.613175] env[65522]: DEBUG nova.network.neutron [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Successfully updated port: 7c90c348-40b6-4412-9db4-4f9bf2e096dc {{(pid=65522) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1623.117566] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "refresh_cache-96e45978-2697-4ac5-a0ba-30e95d7710f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1623.117566] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired lock "refresh_cache-96e45978-2697-4ac5-a0ba-30e95d7710f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1623.117566] env[65522]: DEBUG nova.network.neutron [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Building network info cache for instance {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1623.623539] env[65522]: WARNING openstack [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1623.623539] env[65522]: WARNING openstack [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1623.661697] env[65522]: DEBUG nova.network.neutron [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Instance cache missing network info. {{(pid=65522) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1623.681750] env[65522]: WARNING openstack [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1623.682156] env[65522]: WARNING openstack [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1623.744022] env[65522]: WARNING neutronclient.v2_0.client [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1623.744744] env[65522]: WARNING openstack [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1623.745111] env[65522]: WARNING openstack [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1623.829498] env[65522]: DEBUG nova.network.neutron [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Updating instance_info_cache with network_info: [{"id": "7c90c348-40b6-4412-9db4-4f9bf2e096dc", "address": "fa:16:3e:09:0d:99", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c90c348-40", "ovs_interfaceid": "7c90c348-40b6-4412-9db4-4f9bf2e096dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1624.332457] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Releasing lock "refresh_cache-96e45978-2697-4ac5-a0ba-30e95d7710f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1624.332848] env[65522]: DEBUG nova.compute.manager [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Instance network_info: |[{"id": "7c90c348-40b6-4412-9db4-4f9bf2e096dc", "address": "fa:16:3e:09:0d:99", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c90c348-40", "ovs_interfaceid": "7c90c348-40b6-4412-9db4-4f9bf2e096dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65522) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 1624.333385] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:0d:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '07e9bef1-2b0e-4e4d-997f-de71bb0e213a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7c90c348-40b6-4412-9db4-4f9bf2e096dc', 'vif_model': 'vmxnet3'}] {{(pid=65522) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1624.340807] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1624.341041] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Creating VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1624.341298] env[65522]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dabcf389-10e0-4beb-a5db-dc4d887882a5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1624.361994] env[65522]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1624.361994] env[65522]: value = "task-5115302" [ 1624.361994] env[65522]: _type = "Task" [ 1624.361994] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1624.370154] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115302, 'name': CreateVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1624.557889] env[65522]: DEBUG nova.compute.manager [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Received event network-changed-7c90c348-40b6-4412-9db4-4f9bf2e096dc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1624.558101] env[65522]: DEBUG nova.compute.manager [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Refreshing instance network info cache due to event network-changed-7c90c348-40b6-4412-9db4-4f9bf2e096dc. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1624.558381] env[65522]: DEBUG oslo_concurrency.lockutils [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] Acquiring lock "refresh_cache-96e45978-2697-4ac5-a0ba-30e95d7710f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1624.558534] env[65522]: DEBUG oslo_concurrency.lockutils [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] Acquired lock "refresh_cache-96e45978-2697-4ac5-a0ba-30e95d7710f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1624.558691] env[65522]: DEBUG nova.network.neutron [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Refreshing network info cache for port 7c90c348-40b6-4412-9db4-4f9bf2e096dc {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1624.872183] env[65522]: DEBUG oslo_vmware.api [-] Task: {'id': task-5115302, 'name': CreateVM_Task, 'duration_secs': 0.31203} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1624.872592] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Created VM on the ESX host {{(pid=65522) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1624.872859] env[65522]: WARNING neutronclient.v2_0.client [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1624.873273] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1624.873432] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1624.873743] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1624.874023] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4132138d-2364-4df6-8e34-3e75274c3b95 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1624.880778] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1624.880778] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52128209-7f9b-364d-e919-f928fd179f2b" [ 1624.880778] env[65522]: _type = "Task" [ 1624.880778] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1624.889371] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52128209-7f9b-364d-e919-f928fd179f2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1625.062439] env[65522]: WARNING neutronclient.v2_0.client [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1625.063122] env[65522]: WARNING openstack [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1625.063486] env[65522]: WARNING openstack [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1625.150700] env[65522]: WARNING openstack [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1625.151109] env[65522]: WARNING openstack [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1625.208031] env[65522]: WARNING neutronclient.v2_0.client [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1625.208704] env[65522]: WARNING openstack [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1625.209089] env[65522]: WARNING openstack [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1625.296072] env[65522]: DEBUG nova.network.neutron [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Updated VIF entry in instance network info cache for port 7c90c348-40b6-4412-9db4-4f9bf2e096dc. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1625.296462] env[65522]: DEBUG nova.network.neutron [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Updating instance_info_cache with network_info: [{"id": "7c90c348-40b6-4412-9db4-4f9bf2e096dc", "address": "fa:16:3e:09:0d:99", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c90c348-40", "ovs_interfaceid": "7c90c348-40b6-4412-9db4-4f9bf2e096dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1625.392389] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52128209-7f9b-364d-e919-f928fd179f2b, 'name': SearchDatastore_Task, 'duration_secs': 0.013448} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1625.392639] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1625.392884] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Processing image a889db67-7337-4e32-8e34-642f34402926 {{(pid=65522) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1625.393128] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1625.393274] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1625.393453] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1625.393731] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-45bf6e4a-27c6-416d-b7b2-2083281727d7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1625.403863] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65522) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1625.404060] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65522) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1625.404786] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32055e4a-3968-4d80-a0cc-4b0f68742c66 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1625.410565] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1625.410565] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e1a2b5-c030-d5d1-a8c1-22c8169c316d" [ 1625.410565] env[65522]: _type = "Task" [ 1625.410565] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1625.419705] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e1a2b5-c030-d5d1-a8c1-22c8169c316d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1625.799730] env[65522]: DEBUG oslo_concurrency.lockutils [req-3d27b2b3-dd65-4746-acb1-0f9db737ef69 req-a742246b-1ada-43d3-b7fd-c463499733d1 service nova] Releasing lock "refresh_cache-96e45978-2697-4ac5-a0ba-30e95d7710f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1625.922054] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52e1a2b5-c030-d5d1-a8c1-22c8169c316d, 'name': SearchDatastore_Task, 'duration_secs': 0.011123} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1625.922819] env[65522]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f540608-fc6c-4c41-a6f3-85c88c42aae4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1625.928453] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1625.928453] env[65522]: value = "session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c79cc4-7e70-28c9-0763-f6ec655d4f10" [ 1625.928453] env[65522]: _type = "Task" [ 1625.928453] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1625.937726] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c79cc4-7e70-28c9-0763-f6ec655d4f10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1626.438960] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': session[52c7fdae-e4ca-2112-efbe-02e4b414c314]52c79cc4-7e70-28c9-0763-f6ec655d4f10, 'name': SearchDatastore_Task, 'duration_secs': 0.010078} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1626.439186] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1626.439434] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 96e45978-2697-4ac5-a0ba-30e95d7710f7/96e45978-2697-4ac5-a0ba-30e95d7710f7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1626.439717] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7b47770a-1e95-47bd-a4be-33827be0df61 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1626.446805] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1626.446805] env[65522]: value = "task-5115303" [ 1626.446805] env[65522]: _type = "Task" [ 1626.446805] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1626.455559] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115303, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1626.957425] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115303, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455393} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1626.957775] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a889db67-7337-4e32-8e34-642f34402926/a889db67-7337-4e32-8e34-642f34402926.vmdk to [datastore1] 96e45978-2697-4ac5-a0ba-30e95d7710f7/96e45978-2697-4ac5-a0ba-30e95d7710f7.vmdk {{(pid=65522) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1626.957957] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Extending root virtual disk to 1048576 {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1626.958237] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d283d872-980e-458e-9081-7df17eea5b1d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1626.966242] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1626.966242] env[65522]: value = "task-5115304" [ 1626.966242] env[65522]: _type = "Task" [ 1626.966242] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1626.975970] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115304, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1627.477296] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115304, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065376} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1627.477542] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Extended root virtual disk {{(pid=65522) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1627.478410] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b44349a-bc80-4608-a7ed-a094a368e22d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1627.501347] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Reconfiguring VM instance instance-0000007f to attach disk [datastore1] 96e45978-2697-4ac5-a0ba-30e95d7710f7/96e45978-2697-4ac5-a0ba-30e95d7710f7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1627.501490] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9159e1ad-8669-4332-8dcf-85d494fd3121 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1627.522267] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1627.522267] env[65522]: value = "task-5115305" [ 1627.522267] env[65522]: _type = "Task" [ 1627.522267] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1627.530822] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115305, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1628.033019] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115305, 'name': ReconfigVM_Task, 'duration_secs': 0.275541} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1628.033429] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Reconfigured VM instance instance-0000007f to attach disk [datastore1] 96e45978-2697-4ac5-a0ba-30e95d7710f7/96e45978-2697-4ac5-a0ba-30e95d7710f7.vmdk or device None with type sparse {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1628.033985] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-041d3c23-0a9e-45e9-b5b7-582e9ac9e648 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1628.041614] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1628.041614] env[65522]: value = "task-5115306" [ 1628.041614] env[65522]: _type = "Task" [ 1628.041614] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1628.050795] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115306, 'name': Rename_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1628.551647] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115306, 'name': Rename_Task, 'duration_secs': 0.139466} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1628.551909] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Powering on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1628.552182] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-87891478-5bcf-463d-b9e3-f5ce1d3779ae {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1628.558863] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1628.558863] env[65522]: value = "task-5115307" [ 1628.558863] env[65522]: _type = "Task" [ 1628.558863] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1628.567285] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115307, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1629.069780] env[65522]: DEBUG oslo_vmware.api [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115307, 'name': PowerOnVM_Task, 'duration_secs': 0.435533} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1629.070326] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Powered on the VM {{(pid=65522) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1629.070326] env[65522]: INFO nova.compute.manager [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Took 6.86 seconds to spawn the instance on the hypervisor. [ 1629.070465] env[65522]: DEBUG nova.compute.manager [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Checking state {{(pid=65522) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 1629.071249] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8b34f1-2650-4060-87b7-99375dafb696 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1629.591269] env[65522]: INFO nova.compute.manager [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Took 11.54 seconds to build instance. [ 1630.093630] env[65522]: DEBUG oslo_concurrency.lockutils [None req-e9cf1386-3a3f-4751-a84d-2874e6039b29 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.053s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1630.133086] env[65522]: DEBUG nova.compute.manager [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Received event network-changed-7c90c348-40b6-4412-9db4-4f9bf2e096dc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1630.133253] env[65522]: DEBUG nova.compute.manager [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Refreshing instance network info cache due to event network-changed-7c90c348-40b6-4412-9db4-4f9bf2e096dc. {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11841}} [ 1630.133467] env[65522]: DEBUG oslo_concurrency.lockutils [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] Acquiring lock "refresh_cache-96e45978-2697-4ac5-a0ba-30e95d7710f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1630.133602] env[65522]: DEBUG oslo_concurrency.lockutils [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] Acquired lock "refresh_cache-96e45978-2697-4ac5-a0ba-30e95d7710f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1630.133755] env[65522]: DEBUG nova.network.neutron [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Refreshing network info cache for port 7c90c348-40b6-4412-9db4-4f9bf2e096dc {{(pid=65522) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1630.636738] env[65522]: WARNING neutronclient.v2_0.client [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1630.637924] env[65522]: WARNING openstack [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1630.638291] env[65522]: WARNING openstack [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1630.768716] env[65522]: WARNING openstack [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1630.769146] env[65522]: WARNING openstack [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1630.830240] env[65522]: WARNING neutronclient.v2_0.client [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1630.830884] env[65522]: WARNING openstack [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1630.831246] env[65522]: WARNING openstack [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1630.912680] env[65522]: DEBUG nova.network.neutron [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Updated VIF entry in instance network info cache for port 7c90c348-40b6-4412-9db4-4f9bf2e096dc. {{(pid=65522) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1630.913127] env[65522]: DEBUG nova.network.neutron [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Updating instance_info_cache with network_info: [{"id": "7c90c348-40b6-4412-9db4-4f9bf2e096dc", "address": "fa:16:3e:09:0d:99", "network": {"id": "d7e096c9-9427-4896-a4e7-6889c5914305", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-843273185-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "37a72785dad54301a64694398f167f92", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7c90c348-40", "ovs_interfaceid": "7c90c348-40b6-4412-9db4-4f9bf2e096dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1631.415780] env[65522]: DEBUG oslo_concurrency.lockutils [req-65b0b92e-7587-43f2-b912-6fd00917e802 req-0397aae7-074a-4099-80b4-3066896ebf35 service nova] Releasing lock "refresh_cache-96e45978-2697-4ac5-a0ba-30e95d7710f7" {{(pid=65522) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1652.892081] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1652.892479] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1653.397905] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1653.398133] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1653.398227] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1653.398378] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1653.398522] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1653.398674] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1653.398831] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1653.398960] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1653.901614] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1653.902056] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1653.902125] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1653.902243] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1653.903254] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a35368-f0f0-4f4a-9a97-e3c12b952574 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1653.912481] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04fb0a3-1d0f-4c31-a508-11dc951117a2 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1653.927325] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24fc6d91-bcf0-4b59-8cca-663cf5be7aa1 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1653.934583] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d7daec-ce1c-4963-a588-9df8e2924339 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1653.965967] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180359MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1653.966197] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1653.966374] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1654.992568] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Instance 96e45978-2697-4ac5-a0ba-30e95d7710f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65522) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1654.992986] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1654.992986] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=100GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '1', 'num_vm_active': '1', 'num_task_None': '1', 'num_os_type_None': '1', 'num_proj_37a72785dad54301a64694398f167f92': '1', 'io_workload': '0'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1655.020916] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df99b8c-bf04-4b41-be6e-83df2b52b16b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1655.029257] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35cb49e6-7f6b-4329-8351-aebcc6c5a5fe {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1655.059043] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7da2915-419b-41fd-8582-7ba6108d448b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1655.066911] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a46903e-4873-4a79-91ab-48da606ba8b6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1655.081677] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1655.585074] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1656.089703] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65522) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1656.090199] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.124s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1668.707517] env[65522]: DEBUG oslo_concurrency.lockutils [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1668.707856] env[65522]: DEBUG oslo_concurrency.lockutils [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1669.211892] env[65522]: DEBUG nova.compute.utils [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Using /dev/sd instead of None {{(pid=65522) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1669.715443] env[65522]: DEBUG oslo_concurrency.lockutils [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1670.771323] env[65522]: DEBUG oslo_concurrency.lockutils [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1670.771752] env[65522]: DEBUG oslo_concurrency.lockutils [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1670.771849] env[65522]: INFO nova.compute.manager [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Attaching volume fcaa3670-b09e-4f82-bf96-c8632cc0ebe7 to /dev/sdb [ 1670.801996] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a5ab307-de6e-4be5-bf30-89177302a43a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1670.809653] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a708d22-a66c-4e23-a7e9-6ce6c62dfb91 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1670.823372] env[65522]: DEBUG nova.virt.block_device [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Updating existing volume attachment record: 5fb2edb6-5315-402e-9bc8-d86398aa6ef1 {{(pid=65522) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1675.368926] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Volume attach. Driver type: vmdk {{(pid=65522) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1675.369214] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995007', 'volume_id': 'fcaa3670-b09e-4f82-bf96-c8632cc0ebe7', 'name': 'volume-fcaa3670-b09e-4f82-bf96-c8632cc0ebe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '96e45978-2697-4ac5-a0ba-30e95d7710f7', 'attached_at': '', 'detached_at': '', 'volume_id': 'fcaa3670-b09e-4f82-bf96-c8632cc0ebe7', 'serial': 'fcaa3670-b09e-4f82-bf96-c8632cc0ebe7'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1675.370150] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1839807d-8d19-4947-bf5f-e9ddf80e736f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1675.387811] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-154c03f7-07b8-47ec-a21c-1cc8efda5363 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1675.413896] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Reconfiguring VM instance instance-0000007f to attach disk [datastore1] volume-fcaa3670-b09e-4f82-bf96-c8632cc0ebe7/volume-fcaa3670-b09e-4f82-bf96-c8632cc0ebe7.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1675.414190] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16cda17e-7886-4a12-be91-514ccd2856d0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1675.434147] env[65522]: DEBUG oslo_vmware.api [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1675.434147] env[65522]: value = "task-5115310" [ 1675.434147] env[65522]: _type = "Task" [ 1675.434147] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1675.443734] env[65522]: DEBUG oslo_vmware.api [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115310, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1675.944863] env[65522]: DEBUG oslo_vmware.api [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115310, 'name': ReconfigVM_Task, 'duration_secs': 0.328258} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1675.945134] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Reconfigured VM instance instance-0000007f to attach disk [datastore1] volume-fcaa3670-b09e-4f82-bf96-c8632cc0ebe7/volume-fcaa3670-b09e-4f82-bf96-c8632cc0ebe7.vmdk or device None with type thin {{(pid=65522) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1675.949923] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d357693-78af-488c-a3c2-54d366e4d3df {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1675.966654] env[65522]: DEBUG oslo_vmware.api [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1675.966654] env[65522]: value = "task-5115311" [ 1675.966654] env[65522]: _type = "Task" [ 1675.966654] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1675.976614] env[65522]: DEBUG oslo_vmware.api [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115311, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1676.477638] env[65522]: DEBUG oslo_vmware.api [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1676.979024] env[65522]: DEBUG oslo_vmware.api [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115311, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1677.479624] env[65522]: DEBUG oslo_vmware.api [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115311, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1677.981141] env[65522]: DEBUG oslo_vmware.api [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115311, 'name': ReconfigVM_Task, 'duration_secs': 1.875677} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1677.981386] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995007', 'volume_id': 'fcaa3670-b09e-4f82-bf96-c8632cc0ebe7', 'name': 'volume-fcaa3670-b09e-4f82-bf96-c8632cc0ebe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '96e45978-2697-4ac5-a0ba-30e95d7710f7', 'attached_at': '', 'detached_at': '', 'volume_id': 'fcaa3670-b09e-4f82-bf96-c8632cc0ebe7', 'serial': 'fcaa3670-b09e-4f82-bf96-c8632cc0ebe7'} {{(pid=65522) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1679.016304] env[65522]: DEBUG nova.objects.instance [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lazy-loading 'flavor' on Instance uuid 96e45978-2697-4ac5-a0ba-30e95d7710f7 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1679.224214] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1679.523177] env[65522]: DEBUG oslo_concurrency.lockutils [None req-003718bd-7e28-4ca3-9222-5c8a81e4918d tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.751s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1679.524162] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.300s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1680.026853] env[65522]: INFO nova.compute.manager [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Detaching volume fcaa3670-b09e-4f82-bf96-c8632cc0ebe7 [ 1680.057211] env[65522]: INFO nova.virt.block_device [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Attempting to driver detach volume fcaa3670-b09e-4f82-bf96-c8632cc0ebe7 from mountpoint /dev/sdb [ 1680.057510] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Volume detach. Driver type: vmdk {{(pid=65522) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1680.057700] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995007', 'volume_id': 'fcaa3670-b09e-4f82-bf96-c8632cc0ebe7', 'name': 'volume-fcaa3670-b09e-4f82-bf96-c8632cc0ebe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '96e45978-2697-4ac5-a0ba-30e95d7710f7', 'attached_at': '', 'detached_at': '', 'volume_id': 'fcaa3670-b09e-4f82-bf96-c8632cc0ebe7', 'serial': 'fcaa3670-b09e-4f82-bf96-c8632cc0ebe7'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1680.058703] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a37dad-c3f1-4eaa-977a-23c5fc0f68e9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1680.081705] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f3587e-1289-4779-b97c-51016ea8aba7 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1680.088677] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7860f4-a9ba-4668-9ca8-2aa3b547df45 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1680.110051] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ddfb91-7045-4c7c-a1f4-1188e999c0b0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1680.125898] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] The volume has not been displaced from its original location: [datastore1] volume-fcaa3670-b09e-4f82-bf96-c8632cc0ebe7/volume-fcaa3670-b09e-4f82-bf96-c8632cc0ebe7.vmdk. No consolidation needed. {{(pid=65522) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1680.131169] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Reconfiguring VM instance instance-0000007f to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1680.131492] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74840529-af04-4bf8-ac47-86c0f74d1dee {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1680.151188] env[65522]: DEBUG oslo_vmware.api [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1680.151188] env[65522]: value = "task-5115312" [ 1680.151188] env[65522]: _type = "Task" [ 1680.151188] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1680.160249] env[65522]: DEBUG oslo_vmware.api [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115312, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1680.660689] env[65522]: DEBUG oslo_vmware.api [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115312, 'name': ReconfigVM_Task, 'duration_secs': 0.231745} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1680.660961] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Reconfigured VM instance instance-0000007f to detach disk 2001 {{(pid=65522) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1680.665641] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47f5e427-ecce-40f3-88f1-318100efe42c {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1680.680220] env[65522]: DEBUG oslo_vmware.api [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1680.680220] env[65522]: value = "task-5115313" [ 1680.680220] env[65522]: _type = "Task" [ 1680.680220] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1680.688254] env[65522]: DEBUG oslo_vmware.api [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115313, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1681.190584] env[65522]: DEBUG oslo_vmware.api [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115313, 'name': ReconfigVM_Task, 'duration_secs': 0.131376} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1681.191025] env[65522]: DEBUG nova.virt.vmwareapi.volumeops [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-995007', 'volume_id': 'fcaa3670-b09e-4f82-bf96-c8632cc0ebe7', 'name': 'volume-fcaa3670-b09e-4f82-bf96-c8632cc0ebe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '96e45978-2697-4ac5-a0ba-30e95d7710f7', 'attached_at': '', 'detached_at': '', 'volume_id': 'fcaa3670-b09e-4f82-bf96-c8632cc0ebe7', 'serial': 'fcaa3670-b09e-4f82-bf96-c8632cc0ebe7'} {{(pid=65522) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1681.732914] env[65522]: DEBUG nova.objects.instance [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lazy-loading 'flavor' on Instance uuid 96e45978-2697-4ac5-a0ba-30e95d7710f7 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1682.740885] env[65522]: DEBUG oslo_concurrency.lockutils [None req-1457512b-8d5d-4f43-9adb-6874e5bf1d11 tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.217s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1683.459230] env[65522]: DEBUG oslo_concurrency.lockutils [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1683.459487] env[65522]: DEBUG oslo_concurrency.lockutils [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1683.459702] env[65522]: DEBUG oslo_concurrency.lockutils [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "96e45978-2697-4ac5-a0ba-30e95d7710f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1683.459906] env[65522]: DEBUG oslo_concurrency.lockutils [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "96e45978-2697-4ac5-a0ba-30e95d7710f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1683.460113] env[65522]: DEBUG oslo_concurrency.lockutils [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "96e45978-2697-4ac5-a0ba-30e95d7710f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1683.462451] env[65522]: INFO nova.compute.manager [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Terminating instance [ 1683.966198] env[65522]: DEBUG nova.compute.manager [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Start destroying the instance on the hypervisor. {{(pid=65522) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 1683.966628] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Destroying instance {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1683.967356] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-020e0bd7-d40a-415b-a646-98e3db80291e {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1683.975821] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Powering off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1683.976076] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0793fe1-c874-4f34-892f-c7f17d5cbab0 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1683.982745] env[65522]: DEBUG oslo_vmware.api [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1683.982745] env[65522]: value = "task-5115314" [ 1683.982745] env[65522]: _type = "Task" [ 1683.982745] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1683.991706] env[65522]: DEBUG oslo_vmware.api [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115314, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1684.493572] env[65522]: DEBUG oslo_vmware.api [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115314, 'name': PowerOffVM_Task, 'duration_secs': 0.188481} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1684.493881] env[65522]: DEBUG nova.virt.vmwareapi.vm_util [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Powered off the VM {{(pid=65522) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1684.494077] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Unregistering the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1684.494341] env[65522]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e55ecd43-d7c6-48db-a7e1-e679dee0f5c4 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1684.560080] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Unregistered the VM {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1684.560391] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Deleting contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1684.560657] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Deleting the datastore file [datastore1] 96e45978-2697-4ac5-a0ba-30e95d7710f7 {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1684.560876] env[65522]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0fd32cc3-3132-4aa1-afc8-276ab0ef5323 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1684.567831] env[65522]: DEBUG oslo_vmware.api [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for the task: (returnval){ [ 1684.567831] env[65522]: value = "task-5115316" [ 1684.567831] env[65522]: _type = "Task" [ 1684.567831] env[65522]: } to complete. {{(pid=65522) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1684.575933] env[65522]: DEBUG oslo_vmware.api [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1685.077485] env[65522]: DEBUG oslo_vmware.api [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Task: {'id': task-5115316, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19367} completed successfully. {{(pid=65522) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1685.077919] env[65522]: DEBUG nova.virt.vmwareapi.ds_util [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Deleted the datastore file {{(pid=65522) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1685.077919] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Deleted contents of the VM from datastore datastore1 {{(pid=65522) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1685.078075] env[65522]: DEBUG nova.virt.vmwareapi.vmops [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Instance destroyed {{(pid=65522) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1685.078282] env[65522]: INFO nova.compute.manager [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1685.078535] env[65522]: DEBUG oslo.service.backend._common.loopingcall [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65522) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1685.078727] env[65522]: DEBUG nova.compute.manager [-] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Deallocating network for instance {{(pid=65522) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 1685.078824] env[65522]: DEBUG nova.network.neutron [-] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] deallocate_for_instance() {{(pid=65522) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1685.079089] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1685.079709] env[65522]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1685.079977] env[65522]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1685.131341] env[65522]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1685.590561] env[65522]: DEBUG nova.compute.manager [req-9755fb72-4ec5-4568-b131-2effdcba2dd4 req-ae6453e5-71b0-4158-8c5a-7102f6977eab service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Received event network-vif-deleted-7c90c348-40b6-4412-9db4-4f9bf2e096dc {{(pid=65522) external_instance_event /opt/stack/nova/nova/compute/manager.py:11836}} [ 1685.590703] env[65522]: INFO nova.compute.manager [req-9755fb72-4ec5-4568-b131-2effdcba2dd4 req-ae6453e5-71b0-4158-8c5a-7102f6977eab service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Neutron deleted interface 7c90c348-40b6-4412-9db4-4f9bf2e096dc; detaching it from the instance and deleting it from the info cache [ 1685.590877] env[65522]: DEBUG nova.network.neutron [req-9755fb72-4ec5-4568-b131-2effdcba2dd4 req-ae6453e5-71b0-4158-8c5a-7102f6977eab service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1686.069565] env[65522]: DEBUG nova.network.neutron [-] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Updating instance_info_cache with network_info: [] {{(pid=65522) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1686.093889] env[65522]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fe6f9daa-40e0-48e1-9553-ae6e0d5d67ca {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1686.105764] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c63e34-533c-4b8a-b960-c251943c132a {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1686.134333] env[65522]: DEBUG nova.compute.manager [req-9755fb72-4ec5-4568-b131-2effdcba2dd4 req-ae6453e5-71b0-4158-8c5a-7102f6977eab service nova] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Detach interface failed, port_id=7c90c348-40b6-4412-9db4-4f9bf2e096dc, reason: Instance 96e45978-2697-4ac5-a0ba-30e95d7710f7 could not be found. {{(pid=65522) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11670}} [ 1686.572152] env[65522]: INFO nova.compute.manager [-] [instance: 96e45978-2697-4ac5-a0ba-30e95d7710f7] Took 1.49 seconds to deallocate network for instance. [ 1687.080291] env[65522]: DEBUG oslo_concurrency.lockutils [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1687.080790] env[65522]: DEBUG oslo_concurrency.lockutils [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1687.081196] env[65522]: DEBUG nova.objects.instance [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lazy-loading 'resources' on Instance uuid 96e45978-2697-4ac5-a0ba-30e95d7710f7 {{(pid=65522) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1687.617779] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325d74d5-f399-4475-b1e0-a385a4636216 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1687.625855] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b9d8dc-0e9c-481d-aca5-48c0083822db {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1687.656105] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec5bb073-9147-4167-8180-59d7e5786b1b {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1687.664374] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4458d55-5782-4330-848c-b60647000f88 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1687.677882] env[65522]: DEBUG nova.compute.provider_tree [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1688.181723] env[65522]: DEBUG nova.scheduler.client.report [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1688.687107] env[65522]: DEBUG oslo_concurrency.lockutils [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.606s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1688.710114] env[65522]: INFO nova.scheduler.client.report [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Deleted allocations for instance 96e45978-2697-4ac5-a0ba-30e95d7710f7 [ 1689.220378] env[65522]: DEBUG oslo_concurrency.lockutils [None req-99f7b523-d9ae-4495-9c57-4e6f35820efe tempest-AttachVolumeNegativeTest-1526680072 tempest-AttachVolumeNegativeTest-1526680072-project-member] Lock "96e45978-2697-4ac5-a0ba-30e95d7710f7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.761s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1716.091267] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1716.091816] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1716.091816] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1716.091816] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1716.091967] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1716.092089] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1716.092262] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1716.092400] env[65522]: DEBUG nova.compute.manager [None req-d9c23605-51ad-439f-9979-748648caf427 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65522) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11252}} [ 1716.092544] env[65522]: DEBUG oslo_service.periodic_task [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Running periodic task ComputeManager.update_available_resource {{(pid=65522) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1716.596209] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1716.596525] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1716.596706] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1716.596863] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65522) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1716.597806] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe9cb1d-61ff-4908-99ed-0a455aeecc5f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1716.606804] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5d35be-3e82-4a5c-bb44-073a32ada36d {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1716.621251] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954122bd-a3a9-41de-8dd6-85dff8975ee5 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1716.628424] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb5acaa-1932-4d11-8e72-58d142a1ff28 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1716.659402] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180380MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65522) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1716.659603] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1716.659814] env[65522]: DEBUG oslo_concurrency.lockutils [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65522) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1717.679969] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1717.680264] env[65522]: DEBUG nova.compute.resource_tracker [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=100GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] stats={'failed_builds': '0'} {{(pid=65522) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1717.693563] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0b31c5e-4489-4d2a-bdca-34d613b710a9 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1717.701530] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4878e989-2dd3-40ec-8d78-0660412da8c6 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1717.732684] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50d303f-6235-423f-9fd5-2baf6674920f {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1717.742215] env[65522]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ae53eb-58c6-4575-b476-a33a15d82317 {{(pid=65522) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1717.755692] env[65522]: DEBUG nova.compute.provider_tree [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed in ProviderTree for provider: 9a0d1ac8-ed26-4657-8899-fbceab29db0d {{(pid=65522) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1718.259341] env[65522]: DEBUG nova.scheduler.client.report [None req-d9c23605-51ad-439f-9979-748648caf427 None None] Inventory has not changed for provider 9a0d1ac8-ed26-4657-8899-fbceab29db0d based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65522) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}}